Digital Technologies and Applications: Saad Motahhir Badre Bossoufi Editors
Digital Technologies and Applications: Saad Motahhir Badre Bossoufi Editors
Saad Motahhir
Badre Bossoufi Editors
Digital
Technologies
and Applications
Proceedings of ICDTA 21, Fez, Morocco
Lecture Notes in Networks and Systems
Volume 211
Series Editor
Janusz Kacprzyk, Systems Research Institute, Polish Academy of Sciences,
Warsaw, Poland
Advisory Editors
Fernando Gomide, Department of Computer Engineering and Automation—DCA,
School of Electrical and Computer Engineering—FEEC, University of Campinas—
UNICAMP, São Paulo, Brazil
Okyay Kaynak, Department of Electrical and Electronic Engineering,
Bogazici University, Istanbul, Turkey
Derong Liu, Department of Electrical and Computer Engineering, University
of Illinois at Chicago, Chicago, USA; Institute of Automation, Chinese Academy
of Sciences, Beijing, China
Witold Pedrycz, Department of Electrical and Computer Engineering,
University of Alberta, Alberta, Canada; Systems Research Institute,
Polish Academy of Sciences, Warsaw, Poland
Marios M. Polycarpou, Department of Electrical and Computer Engineering,
KIOS Research Center for Intelligent Systems and Networks, University of Cyprus,
Nicosia, Cyprus
Imre J. Rudas, Óbuda University, Budapest, Hungary
Jun Wang, Department of Computer Science, City University of Hong Kong,
Kowloon, Hong Kong
The series “Lecture Notes in Networks and Systems” publishes the latest
developments in Networks and Systems—quickly, informally and with high quality.
Original research reported in proceedings and post-proceedings represents the core
of LNNS.
Volumes published in LNNS embrace all aspects and subfields of, as well as new
challenges in, Networks and Systems.
The series contains proceedings and edited volumes in systems and networks,
spanning the areas of Cyber-Physical Systems, Autonomous Systems, Sensor
Networks, Control Systems, Energy Systems, Automotive Systems, Biological
Systems, Vehicular Networking and Connected Vehicles, Aerospace Systems,
Automation, Manufacturing, Smart Grids, Nonlinear Systems, Power Systems,
Robotics, Social Systems, Economic Systems and other. Of particular value to both
the contributors and the readership are the short publication timeframe and the
world-wide distribution and exposure which enable both a wide and rapid
dissemination of research output.
The series covers the theory, applications, and perspectives on the state of the art
and future developments relevant to systems and networks, decision making,
control, complex processes and related areas, as embedded in the fields of
interdisciplinary and applied sciences, engineering, computer science, physics,
economics, social, and life sciences, as well as the paradigms and methodologies
behind them.
Indexed by SCOPUS, INSPEC, WTI Frankfurt eG, zbMATH, SCImago.
All books published in the series are submitted for consideration in Web of
Science.
Editors
Digital Technologies
and Applications
Proceedings of ICDTA 21, Fez, Morocco
123
Editors
Saad Motahhir Badre Bossoufi
ENSA Faculty of Sciences
Sidi Mohamed Ben Abdellah University Sidi Mohamed Ben Abdellah University
Fez, Morocco Fez, Morocco
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
We are honored to dedicate the proceedings
of ICDTA’21 to all the participants and
committees of ICDTA’21.
Foreword
It is with deep satisfaction that I write this Foreword to the proceedings of the
ICDTA’21 organized by École Nationale des Sciences Appliquées and Faculté des
sciences which belong to SMBA University, Fez, Morocco, on January 29 and 30,
2021.
This conference was bringing together researchers, academics, and professionals
from all over the world, experts in digital technologies and their applications. This
conference particularly encouraged the interaction of research students and devel-
oping academics with the more established academic community in an informal
setting to present and to discuss new and current work. The papers contributed the
most recent scientific knowledge known in the field of digital technologies and their
applications. Their contributions helped to make the conference as outstanding as it
has been. The organizing and technical program committees put much effort into
ensuring the success of the day-to-day operation of the meeting.
We hope that this book proceedings will further stimulate research in digital
technologies such as artificial intelligence, Internet of things, embedded systems,
network technology, information processing and their applications, in several areas
including hybrid vehicles, renewable energy, robotic, COVID-19, etc.
We feel honored and privileged to serve the best recent developments to you
through this exciting book proceedings.
We thank all authors and participants for their contributions.
Saad Motahhir
Badre Bossoufi
vii
Preface
This conference proceedings volume contains the written versions of most of the
contributions presented during the ICDTA’21. The conference provided a setting
for discussing recent developments in a wide variety of topics including artificial
intelligence, Internet of things, embedded systems, network technology, informa-
tion processing and their applications, in several areas such as hybrid vehicles,
renewable energy, robotic, and COVID-19. The conference has been a good
opportunity for participants from various destinations to present and discuss topics
in their respective research areas.
ICDTA’21 tends to collect the latest research results and applications on digital
technologies and their applications. It includes a selection of 166 papers from 328
papers submitted to the conference from universities and industries all over the
world. All of accepted papers were subjected to strict peer reviewing by 2–4 expert
referees. The papers have been selected for this volume because of quality and the
relevance to the conference.
ICDTA’21 would like to express our sincere appreciation to all authors for their
contributions to this book. We would like to extend our thanks to all the referees for
their constructive comments on all papers; especially, we would like to thank
organizing committee for their hardworking. Finally, we would like to thank the
Springer publications for producing this volume.
Saad Motahhir
Badre Bossoufi
ix
Acknowledgments
We request the pleasure of thanking you for taking part in the first edition of the
International Conference on Digital Technologies and Applications (ICDTA’21).
We are very grateful for your support, so thank you everyone for bringing your
expertise and experience around the conference and engaging in such fruitful,
constructive, and open exchanges throughout the two days of the ICDTA’21.
We would like to extend our deepest thanks and gratitude to all the speakers for
accepting to join us from different countries. Thank you for being such wonderful
persons and speakers. Again, thanks for sharing your insight, knowledge, and
experience.
Of course, this event could not be that successful without the effort of the
organizing and technical program committees. Therefore, Pr. Badre and I would
like to express our sincere appreciation to all of you who generously helped us.
We would like to especially thank all the participants for the confidence and trust
you have placed in our conference. We hope we lived up to your highest
expectations.
Finally, our acknowledgement would be incomplete without thanking the big-
gest source of support. Our deepest gratitude goes to Pr. REDOUANE MRABET,
President of Sidi Mohamed Ben Abdellah University, Prof.
ABDERRAHIM LAHRACH, Director of the National School of Applied Sciences,
Prof. MOHAMMED BELMLIH, Dean of the Faculty of Sciences,
Professor ABDELMAJID SAKA, Deputy Director of the National School of
Applied Sciences, and Professor MOHAMMED ELHASSOUNI, Vice Dean of
Faculty of Sciences. Thank you all for your support and for being all-time open to
hosting such events.
Saad Motahhir
Badre Bossoufi
xi
Contents
xiii
xiv Contents
Network Technology
Explainable Deep Learning Model for COVID-19 Screening in Chest
CT Images . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 981
Mostafa El Habib Daho, Amin Khouani, Mohammed El Amine Lazouni,
and Sidi Ahmed Mahmoudi
Contents xxi
Abstract The inertia weight is one of the inspection parameters that improve the
reliability of the Particle Swarm Optimization (PSO). To define this inertia, several
techniques have been suggested to improve the quality of the PSO algorithm. In this
paper, the aim objective is to determine the new reconfiguration of the radial system
by using two strategies: the chaotic descending inertia weight, and the combination
strategy of the chaotic inertia weight and the success rate technique aiming to improve
the result. To define the reliability and the solidity of these suggested techniques,
IEEE 33 bus is used, and the results are compared with other recent studies where
they have also used the particle swarm optimization by using other strategies to define
the inertia weight parameter. These strategies of inertia weight used in this paper are
selected for their features of speed convergence and precision of solution that is near
to the optimal solution.
1 Introduction
In recent years, several companies have been oriented towards the research about
the new strategies, to improve and optimize energy exploitation by finding suitable
solutions, so these new strategies help to minimize the losses. This paper focuses on
the distribution network’s reconfiguration. Many researchers have proposed different
methods based on mathematical programming for solving problems of losses and
cost minimization.
In this vision, the authors of [1], have used an evolutionary technique using the
particle swarm optimization, to find the optimal solution of restoration of radial power
distribution. The authors of [2], have solved the same problem using the minimum
spanning tree, and in the end, they have compared their solution with the solution of
the authors [1]. For the authors of [3], they have considered that the reactive power is
the weights of lines to reduce losses. On the other hand, the authors of [4] have chosen
the genetic algorithm that provides good quality solutions in reduced execution time.
In this article, i will use an improved and adaptative PSO algorithm, thanks to
its high speed to converge towards an optimal solution, it easy to implement and it
focuses on simple mathematical equations. For testing the solidity and the reliability
of this suggested method, this paper uses the IEEE 33 bus, and compare the results
with other recent studies.
In this vision, the authors of [5], have used in their study, the Particle Swarm
Optimization based on the decreasing inertia weight w to update position and velocity
of the particle then they have used the Backward Forward Sweep to do load flow
analysis. For the authors [6], they have used also the linear decreasing inertia weight
by applying the sigmoid transformation to limit the velocities. On another side, the
authors of [7] have applied the linear decreasing weight by canceling the wend in the
second term. The authors of [8] have tried to solve this problem of reconfiguration
by using a modified shark smell optimization, that based on the same idea of particle
swarm optimization, and he concludes that this method solves the problem in a
significant time and help to improve the profile voltage at each node.
Considering the authors of [9], where they have made a comparative study between
the different strategies used to calculate the inertia weight parameter, and to validate
the performance of their study they based on five famous mathematic equations,
at the end of their study, they have concluded that the chaotic inertia weight is
the best strategy for better accuracy, however, the random inertia weight strategy
is best for better efficiency. On the other hand, the authors of [10] have based on
the chaotic inertia weight by combined the swarm success rate parameter with the
chaotic mapping to set the new inertia weight coefficient, aiming to verify the quality
and the efficiency of this proposed method, they tested their strategy by searching the
solution of these function (Griewank, Rastrigin, Rosenbrock, Schaffer f6, Sphere),
and they have concluded that the swarm success rate is a useful tool to improve the
reliability of any swarm focused on optimization algorithms.
Hence, in this paper, the aim objective is to adjust the inertia weight parameter by
using the two tools described in the previous paragraph (the chaotic inertia weight and
the swarm success rate combined with the chaotic mapping). To study this problem,
we have divided the article into four sections. Section two presents the objective
function and define the constraints. Describe the algorithm of the chosen method
and present the case study. Section three discusses and analyses the results with the
other recent works. And in the five and the final section, i conclude the research and
i determine the future works.
New Reconfiguration of the Radial Distribution Network … 5
2 Problematic Reformulation
As described in the previous section, the peak demand period is when all resources
are operating at maximum, and among the drawback of this period is that give rise to
unnecessary expenses for the electric companies resulting from losses. As the load
increases, these losses are more significant. Aiming to reduce the losses, this paper is
focused on reconfiguration of the radial system by using two kinds of inertia weight
to implement the PSO algorithm.
In this paper, to calculate losses, it seems important to use the following expression:
Pj = Rk ∗ Ik 2 (1)
k∈S
With S is the set of the network’s lines and Ik is the current at line k, Rk is the
current at line k.
This problem of the loss’s optimization solved under the following constraints
[11]:
Kirchhoff’s law:
I∗X=0 (2)
Where; I: row vector of current and X: incidence matrix of graph (Xij = 0 if there
are no arcs between i and j; Xij = 1 else);
Tolerance limit:
Vin nominal voltage at node i, Vi is the voltage at node i, and εimax is tolerance
limit at node i [4] (±5% for HTA and +6%/−10%BT).
Admissible current constraint:
Ik ≤ Ik,maxadm (4)
Ik : current of the line k and Ik,maxadm : current maximum admissible of the line k.
Radial topology constraint:
For the reasons to have a simple, inexpensive operation and keeping the security
and protection of distribution power grid, the radial configuration is requested. It
means that each loop should have an open line. To check this topology, the following
constraints should be considered:
Total number of main loops:
6 M. M’dioud et al.
Where Nbr is the total branches of the network, Nbus is the total number of buses
and Nmain loops are the total number of loops in the network.
The Total number of sectionalizing switches
The total number of open switches must be equal to the number of main loops in
the network.
To solve this problem, it seems useful to break up this problem into two important
elements, the first element concerning the reconfiguration of the radial distribution
system using the PSO algorithm and the second element concerning the Newton-
Raphson methods to apply the power flow module to check the set of the constraints.
This load flow method is chosen due to its feature of a fast convergence rate [12].
n
Succrate = Successti /n (9)
i=1
According to the choice of the inertia weight calculation strategy, adjust and
calculate the inertia coefficient by using [13]:
After adjusting the inertia weight by selecting one of the previous equations that
define this parameter:
Update the velocity by using this equation [13]:
Define the new fitness function value for the new position [13]:
Pbestti f xit+1 > Pbestit
Pbestt+1 = (14)
t
xit+1 f xit+1 ≤ Pbestit
Step 8: until iter = maxiter, go to step 4. Else print the optimal results.
Step 9: display result.
In IEEE graph has 33 buses. There are 37 branches, 32 close switches, and 5 open
switches. The red line indicates the tie-line switches. In the initial case, we assume
that {33-34-35-36-37} is the set of the opened lines. The IEEE 33 features (electrical
and topological) are given in the reference [14], the following figure shows the
architecture of the network at the initial case (Fig. 1):
The respective loops, hence, dimensions, the search space for respective dimen-
sion and the parameters of PSO are given in [8].
8 M. M’dioud et al.
After the implementation of the IEEE 33 bus data in MATLAB, we obtained the new
radial distribution that presented in the next schema (Figs. 2 and 3):
In this case, and after reconfiguration by using the chaotic inertia weight, it’s
clear that the new tie lines are 7-9-14-32-37. For the second suggested method using
the combination of the success rate and the chaotic inertia weight, the tie lines are
7-11-14-28-36. To define the updated loss value of the new distribution network
and improve the voltage profile at each node, the Newton Raphson method [12]
(applicated by using Matpower version 4.1). The next table summarizes the result of
New Reconfiguration of the Radial Distribution Network … 9
Fig. 3 Network distribution after reconfiguration by using PSO with combination of the success
rate and the chaotic inertia weight
this paper and gives the result of other studies to make difference and conclude the
feature and the benefits of this study.
As presented in the above table, concerning the tie switch, for the case of recon-
figuration by using the chaotic inertia weight, the new tie lines are 7-9-14-32-37,
while the new tie lines in the case of using the combination of success rate and the
chaotic inertia are 7-11-14-28-36. When we talk about the role of these strategies to
reduce losses, we notice that the losses in the case of PSO using the chaotic inertia
weight equal to 138.9276 kW, and in the case of the using a combination of success
rate and the chaotic inertia weight the losses become 143.9227 kW. These values are
lesser than the losses in the base case (208.459 kW) and the case [5] where the value
of losses is 148.3075 kW. But they are greater than the value found by [7] which
equals 136.3 kW.
For the voltage profile side, the Table 1 proves that in the case of PSO using the
chaotic inertia weight, the minimum voltage profile equal to 0.9423 p.u at node 32,
and in the case of PSO using the combination of success rate and the chaotic inertia
weight the losses equal to 0.93682 p.u at node 18.these values are much improved
10 M. M’dioud et al.
Fig. 4 Voltage profile before and after the reconfiguration using chaotic inertia weight
than the base case 0.91075 p.u at bus 18, and also improved than [5] where the
minimum voltage profile is 0.9395 p.u at bus 32. Also, this value found by using the
chaotic inertia weight is perfect than [7], where the minimum voltage in this study
equal to 0.940 p.u at node 22.
To conclude, it seems clear that the PSO using the chaotic inertia weight gives the
best result than the PSO using the combination of the success rate and the chaotic
inertia weight. This first proposed strategy gives the same result as [8]. However, this
first suggested method reduces the losses with voltage profile improved than [7].
The following figure shows the difference between the voltage profile in the case
of the chaotic inertia weight method and in the base case (Fig. 4):
The Fig. 5 shows the difference between the voltage profile in the case of the
combination method and in the base case. In addition, it is important to note that this
first proposed algorithm takes 46.68s and the second one takes about 55.178566 s
for that they run.
New Reconfiguration of the Radial Distribution Network … 11
Fig. 5 Voltage profile before and after the reconfiguration using the combination of success rate
and the chaotic inertia weight
4 Conclusion
This paper is concerned with a study of two kinds of strategies to determine the
inertia weight, the chaotic inertia weight, and the combination strategy. The main
objective of this research is to define a new reconfiguration of the radial distribution,
aiming to reduce losses and improve voltage profile.
To check the reliability and the performance of these strategies, we use IEEE
33 bus, to test these strategies. The results found are compared with another recent
research focused on the PSO algorithm. According to the results, it’s clear that using
the chaotic inertia weight help to reduce losses better than the combination strategy
of the success rate and the chaotic inertia weight. In the framework of the profile
voltage improvement, these two strategies found the same result. But the combination
strategy takes more time than the chaotic inertia weight strategy to execute.
Finally, the chaotic inertia weight is the best strategy for better reconfiguration
and improved losses. In the next study, I will use this perfect strategy to find the best
allocation and sizing of a distributed generation.
References
1. Oliveira LW, Oliveira EJ, Silva IC, Gomes FV, Borges TT, Marcato AL, Oliveira ÂR (2015)
Optimal restoration of power distribution system through particle swarm optimization. In: IEEE
Eindhoven PowerTech, p 13
12 M. M’dioud et al.
2. Mohamad H (2019) Power system restoration in distribution network using kruskal’s algorithm.
Indonesian J Electr Eng Comput Sci 16(1):8
3. Sudhakar TD (2017) Loss minimization in distribution network by using Prim’s algorithm. Int
J Appl Eng Res 9(24):15
4. Tomoiagă B, Chindriş M, Sumper A, Villafafila-Robles R, Sudria-Andreu A (2013) Distribution
system reconfiguration using genetic algorithm based on connected graphs. Electric Power Syst
Res J 104(216–225):10
5. Sudhakara Reddy AV, Damodar Reddy M (2016) Optimization of network reconfiguration by
using particle swarm optimization. In: 1st IEEE international conference on power electronics,
intelligent control and energy systems (ICPEICES)
6. Tandon A, Saxena D (2014) Optimal reconfiguration of electrical distribution network using
selective particle swarm optimization algorithm
7. Atteya II, Ashour H, Fahmi N, Strickland D (2017) Radial distribution network reconfiguration
for power losses reduction using a modified particle swarm optimisation
8. Juma SA (2018) Optimal radial distribution network reconfiguration using modified shark smell
optimization. Pan African University Institute for Basic Sciences Technology and Innovation
9. Bansal JC, Singh PK, Saraswat M, Verma A, Jadon SS, Abraham A (2011) Inertia weight
strategies in particle swarm optimization
10. Arasomwan MA, Adewumi AO (2013) On adaptive chaotic inertia weights in particle swarm
optimization
11. Enacheanu FB (2008) outils d’aide à la conduite pour les opérateurs des réseaux. https://tel.
archives-ouvertes.fr/tel-00245652
12. Sharma A, Saini M, Ahmed M (2017) Power flow analysis using NR method. In: International
conference on innovative research in science, technology and management
13. Kennedy J, Eberhart R (1995) Particle swarm optimization. In: IEEE international conference
on neural networks, proceedings, vol 4
14. Baran ME, Wu FF (1989) Network reconfiguration in distribution systems for loss reduction
and load balancing. EEE Trans Power Deliv 4(2):7
Convolutional Neuronal Networks
for Tumor Regions Detection
in Histopathology Images
Abstract Recent advances in deep learning and image sensors present the oppor-
tunity to unify these two complementary fields of research for a better resolution of
the image classification problem. Deep learning provides image processing with the
representational power necessary to improve the performance of image classification
methods. The deep networks are considered as the most powerful tool in visual recog-
nition, notably on the automated analysis of histological slides using Whole Slide
Image (WSI) scanning techniques. In this paper, we proposed a comparative study
analysis for efficient localization of tumor regions in kidney histopathology WSIs
by employing state-of-the-art convolutional deep learning networks such as ResNet
and VGGNet which can be fine-tuned to the target problem and exploit the powerful
transfer learning techniques that save low computational costs. Experimental results
conducted on an independent set of 17.300 images derived from 12 WSIs show
that ResNet50 and VGG19 combined with VGG16 achieve almost 97% and 96% of
accuracy respectively, for classifying tumor and normal tissues to identify relevant
regions of interest (ROIs).
M. L. Benomar (B)
University of Ain Temouchent Belhadj Bouchaib, Ain Temouchent, Algeria
e-mail: [email protected]
M. L. Benomar · N. Settouti
Biomedical Engineering Laboratory, University of Tlemcen, Tlemcen, Algeria
e-mail: [email protected]
R. Xiao · X. Descombes
Inria, Sophia Antipolis, Biot, France
e-mail: [email protected]
X. Descombes
e-mail: [email protected]
D. Ambrosetti
Laboratoire Central d’Anatomo Pathologie, CHU Nice/Hôpital Pasteur, Nice, France
e-mail: [email protected]
1 Introduction
2 Related Works
Research on deep learning through convolutional and other neural networks has paved
the way for a representation machine learning approach based on an unsupervised
extraction of descriptors (low, medium and high level), i.e. independent of any human
intervention that could affect these performances. In recent years, several papers have
been published showing the potential of deep learning in digital histopathology.
According to the review of state-of-the-art deep learning approaches that have been
used in the context of histopathological image analysis presented by Srinidhi et al.
[1], among the supervised learning techniques, there are three major canonical deep
learning models based on the nature of tasks that are solved in digital histopathology:
classification, regression and segmentation-based models.
Most deep learning methods applied in digital pathology have been successfully
evaluated in renal clear cell carcinoma [26], breast cancer [11], cervical cancer
[16], prostate cancer [27], where several multiple publicly databases are available.
However, there is a need for a relevant one which is more significant and represen-
tative of real clinical task. In this paper, we proposed a histopathological Kidney
database, it includes 17.300 images issue from 12 WSIs Hematoxylin and eosin
(H&E) stained slides of Kidney tumor collected at the University Hospital Center of
Nice (France).
Whilst large part of the literature deals with tumor regions segmentation [28, 29]
or cell and nuclei morphology analysis [26, 30]. Meanwhile, [31] investigate the
vascular network on the same dataset study in this paper, in scanning the network
irregularity indicating the severity of the cancer i.e. malignant tumor and inversely,
the regularity of the network for low grade tumor.
In our case, we are particularly interested in the highest level of magnification,
and in the segmentation of tumor regions, although not requiring the intervention of
an expert, this task could gain in precision to become sufficiently reliable and thus
avoid having use of manual segmentation. We propose to study two deep learning
architectures, VGGNet and ResNeT which their weight configuration is publicly
16 M. L. Benomar et al.
available and has been used in many other applications and challenges as a basic
functionality extractor. However, these two architectures include several parameters,
which can be a bit difficult to manage and interesting to study. Indeed, VGGNet 16 &
19 consists of 16 and 19 convolutional layers respectively and are very attractive due
to their very uniform architecture. Like the canonical approach of LeNet and AlexNet,
with only 3 × 3 convolutions, but lots of filters. It is currently the most popular choice
in the community for extracting functionality from images. The ResNet residual
neural network is characterized by an architecture with “jump connections” and
presents a very heavy batch normalization. These jump connections are also called
gated 15 units or gated recurrent units and have a strong similarity to recent successful
elements applied in RNNs.
3 Proposed Approach
In this paper, we evaluate different deep learning networks to deal with histopathology
images task that may assist pathologists in locating tumor regions and allowing
further complex image analysis exams to be performed only on relevant regions
within the entire slide image (WSI) as can be seen from Fig. 1, thereby achieving
more computational efficiency increase pathologists throughput.
To develop this system for identifying tumor tissues on histopathological images,
our approach focused on data augmentation and the implementation of ResNet and
a combination of VGG19 & VGG16. To perform data augmentation, the whole
slide image was divided into patches by cropping (500 × 500) regions from the high
resolution image to increase the amounts of training images [32, 33], where the goal is
Fig. 1 Tumor Regions Detection Pipeline. 1 Tissue whole slide images (WSI); 2 Extracted patches
from whole slide images; 3 Feeding the Deep Network with the Normal and Cancerous slides
patches extracted by sliding window; 4 Patches classification as Tumor or Not-tumor; 5 Binary
mask generated where each pixel of the map corresponds to a (500 × 500 pixels) patch in the
input WSI; 6 Small patches removed using morphological operations. The resulting image is then
displayed by sliding window to obtain the final tumor map for the entire whole slide image
Convolutional Neuronal Networks for Tumor Regions Detection … 17
to classify the WSI patches as belonging to tumor regions (malignant) or normal tissue
(benign class). Additionally, we applied affine (low-level) transformations to modify
images without biasing the classification by using more morphological operators.
Besides the data augmentation, computational burden and memory load are also
impacted; consequently, each patch (sub-image) was processed independently.
The major strength of ResNet architecture lies in its ability to train success-
fully very deep neural networks with over 150 layers, and addresses the problem of
vanishing gradients (i.e. the repeated multiplication can make the gradient extremely
small when back-propagating to earlier layers) by providing shortcuts connection
at each layer of the network allowing the gradient to pass through the layers more
easily [12]. As discussed, we investigate the way of applying popular models such as
the VGG network [13] due to its simplicity and success in the ImageNet challenge.
VGG network uses small kernels (only 3 × 3) in every convolution layer in order
to extract fine features present in images. Additionally, a significant improvement
can be achieved by pushing the depth to 16–19 weight layers. Therefore, to correctly
classify tumor and other tissue regions from histopathology whole slide images we
combine both VGG16 and VGG19 to improve the performance. Indeed, combining
different feature extraction methods increases the classification accuracy [34].
Fig. 2 Histopathologic patches examples resize to 224 × 224 pixels at 400× magnification
did not improve for a certain number of epochs (3 epochs). The tests for patch
classification were implemented using Keras library and trained in 25 and 29 min
for ResNet50 and VGG19 & VGG16 respectively, on Colab (Google Colaboratory)
GPU environment.
4.3 Results
The two deep networks described in the previous section were evaluated on the
histopathology whole slide images dataset and the experiments were carried out on
the training, validation and test subset, where we apply the exact same preprocessing
by rescaling images before evaluating the network in order to reach maximum perfor-
mance. The number of trainable parameters for each model used are 40.666.881 for
ResNet50 and 34.740.097 for VGG19 and VGG16.
Figures 3 and 4 present the training and validation set loss of ResNet50 and VGGs
models respectively, the variance is low ensuring that our models are not over-fitted.
Additionally, the accuracy results show that both ResNet50 and VGG19 & VGG16
models perform well and correctly classify cancerous regions. However, ResNet50
achieved slightly better performance than VGG19 & VGG16 models, as shown in
Figs. 3 and 4 we plotted the Receiver Operating Characteristic (ROC) Curve of our
models, where it achieved an area under the curve (AUC) equal to 0.972 and 0.967
respectively.
These results are in good agreement with the ResNet50 model which use the
transfer learning techniques that can improve the classification performance with a
limited training dataset. Moreover, Deep Residual Network uses bottleneck residual
Fig. 5 ROIs identification results (from left to right): input image, ground truth image, ResNet50
and VGG19&VGG16 result respectively
block design to mitigate the vanishing gradient problem by allowing alternate shortcut
path for gradient to flow through.
The fusion from VGG19 and VGG16 convolutional layers also achieved impor-
tant improvements by incrementally adapting the pre-trained features and correlate
the sizes of the receptive fields of the convolutional layers with the basic image
elements (nucleus, vessels or a particular tissue pattern) to increase the recogni-
tion accuracy on tumor vs not-tumor regions classification in histopathology images
dataset. Besides the training and validation graphs, the visual results are also impor-
tant as an evaluation tool. As shown in Fig. 5, the two networks provide important
precision in locating the tumor regions for a given whole slide image and assure
that no critical region is overlooked during diagnosis. Subsequently, for future work
some tumor characteristics must be indicated in order to retain particular regions,
since the model provides tumor regions not delimited by the physician in the ground
truth.
5 Conclusion
In this study, we compare ResNet and VGG convolutional deep learning models that
can identify RCC (renal cell carcinoma) regions from normal tissue in histopathology
whole slide images. The CNN’s successfully classify the cancerous regions where the
ResNet50 achieves the best performance accuracy by adding shortcuts connections
to address the accuracy degradation caused by gradient vanishing effect. Moreover,
transfer learning and data augmentation can be considered to overcome the lack of
data in medical field. As a future work, a post-processing step is necessary to obtain
tumor probability maps for the whole slide image. Also, we need to visualize the
overall information captured by the networks and the critical learned features in
differentiation of cancer regions to further improve the results by adding more data.
This opens the possibility to perform classification of the extracted regions into RCC
subtypes.
References
1. Srinidhi CL, Ciga O, Martel AL (2019) Deep neural network models for computational
histopathology: a survey. arXiv preprint arXiv:1912.12378
2. Verma R, Kumar N, Patil A, Kurian NC, Rane S, Sethi A (2020) Multi-organ nuclei
segmentation and classification challenge 2020
3. Aresta G, Araujo T, Kwok S, Chennamsetty SS, Safwan M, Varghese A, Marami B, Prastawa
M, Chan M, Donovan MJ, Fernandez G, Zeineh J, Kohl M, Walz C, Ludwig F, Braunewell S,
Baust M, Vu QD, To MNN, Kim E, Kwak JT, Galal S, Sanchez-Freire V, Brancati N, Frucci
M, Riccio D, Wang Y, Sun L, Ma K, Fang J, Koné I, Boulmane L, Campilho A, Eloy C,
Polonia A, Aguiar P (2018) BACH: grand challenge on breast cancer histology images. CoRR
abs/1808.04277
4. Borovec J, Kybic J, Arganda-Carreras I, Sorokin DV, Bueno G, Khvostikov AV, Bakas S, Chang
EI, Heldmann S, Kartasalo K, Latonen L, Lotz J, Noga M, Pati S, Punithakumar K, Ruusuvuori
P, Skalski A, Tahmasebi N, Valkonen M, Venet L, Wang Y, Weiss N, Wodzinski M, Xiang Y,
Xu Y, Yan Y, Yushkevic P, Zhao S, Munoz-Barrutia A (2020) ANHIR: automatic non-rigid
histological image registration challenge. IEEE Trans Med Imaging 1 (2020)
5. Bandi P, Geessink O, Manson Q, Dijk MV, Balkenhol M, Hermsen M, Bejnordi BE, Lee B,
Paeng K, Zhong A, Li Q, Zanjani FG, Zinger S, Fukuta K, Komura D, Ovtcharov V, Cheng S,
Zeng S, Thagaard J, Dahl AB, Lin H, Chen H, Jacobsson L, Hedlund M, Cetin M, Halici E,
Jackson H, Chen R, Both F, Franke J, Kusters-Vandevelde H, Vreuls W, Bult P, van Ginneken
B, van der Laak J, Litjens G (2019) From detection of individual metastases to classification
of lymph node status at the patient level: the camelyon17 challenge. IEEE Trans Med Imaging
38(2):550–560
6. Ehteshami Bejnordi B, Veta M, Johannes van Diest P, van Ginneken B, Karssemeijer N, Litjens
G, van der Laak JAWM, the CAMELYON16 Consortium (2017) Diagnostic assessment of deep
learning algorithms for detection of lymph node metastases in women with breast cancer. JAMA
318(22), 2199–2210
7. Veta M, Heng YJ, Stathonikos N, Bejnordi BE, Beca F, Wollmann T, Rohr K, Shah MA,
Wang D, Rousson M, Hedlund M, Tellez D, Ciompi F, Zerhouni E, Lanyi D, Viana MP,
Kovalev V, Liauchuk V, Phoulady HA, Qaiser T, Graham S, Rajpoot NM, Sjoblom E, Molin J,
Paeng K, Hwang S, Park S, Jia Z, Chang EI, Xu Y, Beck AH, van Diest PJ, Pluim JPW (2018)
Predicting breast tumor proliferation from whole-slide images: the TUPAC16 challenge. CoRR
abs/1807.08284
8. Qaiser T, Rajpoot NM (2019) Learning where to see: a novel attention model for automated
immunohistochemical scoring. IEEE Trans Med Imaging 38:2620–2631
9. Serag A, Ion-Margineanu A, Qureshi H, McMillan R, Saint Martin MJ, Diamond J, O’Reilly
P, Hamilton P (2019) Translational ai and deep learning in diagnostic pathology. Front Med
6:185
10. Dimitriou N, Arandjelovi O, Caie PD (2019) Deep learning for whole slide image analysis: an
overview. Front Med 6:264
11. Singh OV, Choudhary P, Thongam K (2020) A study on deep learning for breast cancer detection
in histopathological images. In: Nain N, Vipparthi SK, Raman B (eds) Computer vision and
image processing. Springer, Singapore, pp 36–48
12. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: 2016
IEEE conference on computer vision and pattern recognition, CVPR 2016, Las Vegas, NV,
USA, 27–30 June 2016. IEEE Computer Society, pp 770–778
22 M. L. Benomar et al.
13. Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image
recognition. In: Bengio Y, LeCun Y (eds) 3rd International conference on learning representa-
tions, ICLR 2015, San Diego, CA, USA, 7–9 May 2015, Conference Track Proceedings
14. Albarqouni S, Baur C, Achilles F, Belagiannis V, Demirci S, Navab N (2016) Aggnet: deep
learning from crowds for mitosis detection in breast cancer histology images. IEEE Trans Med
Imaging 1313–1321
15. Valkonen M, Isola J, Ylinen O, Muhonen V, Saxlin A, Tolonen T, Nykter M, Ruusuvuori P
(2020) Cytokeratin-supervised deep learning for automatic recognition of epithelial cells in
breast cancers stained for ER, PR, and Ki-67. IEEE Trans Med Imaging 39(2):534–542
16. Halicek M, Shahedi M, Little JV, Chen AY, Myers LL, Sumer BD, Fei B (2019) Head and neck
cancer detection in digitized whole-slide histology using convolutional neural networks. Sci
Rep 9(1):14043
17. Zhang Z, Chen P, McGough M, Xing F, Wang C, Bui M, Xie Y, Sapkota M, Cui L, Dhillon J,
Ahmad N, Khalil FK, Dickinson SI, Shi X, Liu F, Su H, Cai J, Yang L (2019) Pathologist-level
interpretable whole-slide cancer diagnosis with deep learning. Nat Mach Intell 1(5):236–245
18. Xie Y, Xing F, Shi X, Kong X, Su H, Yang L (2018) Efficient and robust cell detection: a
structured regression approach. Med Image Anal 44:245–254
19. Xie Y, Kong X, Xing F, Liu F, Su H, Yang L (2015) Deep voting: a robust approach toward
nucleus localization in microscopy images. In: Navab N, Hornegger J, Wells WM, Frangi AF
(eds) Medical image computing and computer-assisted intervention - MICCAI 2015. Springer,
Cham, pp 374–382
20. Sirinukunwattana K, Raza SEA, Tsang Y, Snead DRJ, Cree IA, Rajpoot NM (2016) Locality
sensitive deep learning for detection and classification of nuclei in routine colon cancer
histology images. IEEE Trans Med Imaging 35(5):1196–1206
21. Naylor P, Lae M, Reyal F, Walter T (2019) Segmentation of nuclei in histopathology images
by deep regression of the distance map. IEEE Trans Med Imaging 38(2):448–459
22. Xing F, Cornish TC, Bennett T, Ghosh D, Yang L (2019) Pixel-to-pixel learning with weak
supervision for single-stage nucleus recognition in Ki67 images. IEEE Trans Biomed Eng
66(11):3088–3097
23. Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation.
In: 2015 IEEE conference on computer vision and pattern recognition (CVPR), pp 3431–3440
24. Chan L, Hosseini MS, Rowsell C, Plataniotis KN, Damaskinos S (2019) Histosegnet: semantic
segmentation of histological tissue type in whole slide images. In: The IEEE international
conference on computer vision (ICCV), October 2019
25. Qu H, Riedlinger G, Wu P, Huang Q, Yi J, De S, Metaxas D (2019) Joint segmentation and
fine-grained classification of nuclei in histopathology images. In: 2019 IEEE 16th international
symposium on biomedical imaging (ISBI 2019), pp 900–904
26. Tian K, Rubadue CA, Lin DI, Veta M, Pyle ME, Irshad H, Heng YJ (2019) Automated clear
cell renal carcinoma grade classification with prognostic significance. PLoS ONE 14(10):1–16
27. Nagpal K, Foote D, Liu Y, Chen C, Wulczyn E, Tan F, Olson N, Smith J, Mohtashamian A,
Wren JH, Corrado G, MacDonald R, Peng L, Amin M, Evans A, Sangoi A, Mermel C, Hipp
J, Stumpe M (2019) Development and validation of a deep learning algorithm for improving
Gleason scoring of prostate cancer. Nature Partner J (NPJ) Digit Med 2:48
28. Cheng J, Mo X, Wang X, Parwani A, Feng Q, Huang K (2017) Identification of topological
features in renal tumor microenvironment associated with patient survival. Bioinformatics
34(6):1024–1030
29. de Bel T, Hermsen M, Litjens G, van der Laak J (2018) Structure instance segmentation in
renal tissue: a case study on tubular immune cell detection. In: Stoyanov D, Taylor Z, Ciompi
F, Xu Y, Martel A, Maier-Hein L, Rajpoot N, van der Laak J, Veta M, McKenna S, Snead D,
Trucco E, Garvin MK, Chen XJ, Bogunovic H (eds) Computational pathology and ophthalmic
medical image analysis. Springer, Cham, pp 112–119
30. Tabibu S, Vinod PK, Jawahar CV (2019) Pan-renal cell carcinoma classification and survival
prediction from histopathology images using deep learning. Sci Rep 9(1):10509
Convolutional Neuronal Networks for Tumor Regions Detection … 23
1 Introduction
Plant diseases cause considerable crops damage, which greatly reduces produc-
tion. Several types of diseases are destructive [1] especially for the cereals that
are the subject of our study. In Morocco, wheat is the most consumed cereal, its
consumption is estimated at 173 kg/year/person with a production in 2016 of 2.7
Mt [2]. However, this production is affected by many biotic constraints; including
cryptogamic diseases. Septoria and brown rust caused by Zymoseptoria tritici and
Puccinia triticina, respectively, are our main wheat diseases [3]. The symptoms of
these two leaf diseases are specific, they cause significant losses [4]. In order to avoid
agricultural yield damage, protecting wheat from disease is essential to guarantee
the quality and quantity of crops [5]. An effective protection strategy must begin
with early detection of the disease in order to choose the appropriate treatment at the
right time to prevent its spread [6]. Usually, in the absence of an agricultural warning
system, the diagnosis of diseases is done in two stages. The first consists of peri-
odic inspections to detect anomalies in plant growth. The second consists in taking
samples of diseased plants which will be analyzed in the laboratory by phytopathol-
ogists [1]. For farmer, continuous monitoring is difficult and time consuming. These
limitations make important the automation of an early diagnosis diseases for crop
protection [5].
In this context, the image processing by machine learning approaches provides
quick and easily usable solutions. For disease diagnosis, the classifying images by
machine learning are divided in two classical methods and deep forms. Classification
methods based on classical learning use a limited number of images [7]. They allow
iteration more quickly and therefore open the possibility of trying several techniques
in a shorter time [8].These algorithms are easier to interpret and understand [9].
However, their accuracy are lower than those of deep learning methods, in particular
the CNN method. The introduction of deep learning techniques in agriculture, espe-
cially in phytopathology field [10], has only started to take place in the last six years,
to a rather limited extent. The author in reference [11] presented in their research
the importance and potentialities of deep learning in agriculture and plant disease.
The research [10] presented a new method for identifying rice diseases based on
CNN techniques, using a data set of 500 natural images of healthy and diseased rice
leaves. Reference [12] used deep learning architectures for the diagnosis of various
plant diseases. Experimental results of [13] demonstrate that VGG-FCN-VD16 and
VGG-FCN-S system outperforms conventional CNN architectures for wheat disease
recognition. In this work, we try to experiment the performance of new CNN architec-
tures in Moroccan wheat disease context using our specific image and by comparing
results of small and large database.
2.1 Data
Data acquisition is the first preparing step. Data can be acquired from the web, using
a database, or manually by taking pictures with RGB cameras. They are prepared and
inoculated in plant protection laboratory of the Agronomic and Veterinary Hassan
II Institute. We have taken pictures using smartphone camera with 27 Mega Pixels.
Three diseases are chosen: brown rust, Septoria leaf spot and powdery mildew. We
obtained 100 images for each disease. Figure 1 shows an example of the images.
Early Detection of Wheat Diseases in Morocco … 27
Fig. 1 Sample images of affected wheat leaves: a Healthy, b Septoria, c Powdery mildew, d Brown
rust
For training and testing our deep learning algorithm, it is important to separate the
images. The goal is to keep some of the data outside of training, which will be used to
test the reliability of the model. The most popular separation is 80% for training data
and 20% for test data. This is to make sure that the model can work in unexpected
cases.
2.3 Methodology
In this work, we expose the adopted methodology whose main objective is the
construction of early detection model for wheat diseases diagnosis. Our approach
begins by the image processing. Next, we choose the most suitable model architec-
ture for the recognition of wheat diseases. We will also study the effect of hyper
parameters variations and their influences on the model. To finish, we present and
discuss the obtained results.
Image Processing
Image Labeling
28 K. Aitelkadi et al.
Image labeling consists of associating the images with labels, for example associating
the image of wheat disease with its label which is the name of this disease. These
labels are a knowledge base for learning the model. The purpose of the model will
be to associate the appropriate label with each test image. So you have to make sure
you create correct labels.
Image Transformation
Neural networks perform better when the input data has common information, the
same background and the same image size. Our images are characterized by the
variation of the backgrounds. Following this variation we decided to proceed with
segmentation to homogenize the background, this procedure is explained in Fig. 2.
The segmentation was done using OpenCV graphics library. We started by
converting the RGB image to an HSV format. Subsequently, we applied an edge
detection algorithm. Finally we created a mask to fill the detected part. Figure 3
shows an image before and after segmentation.
Wheat Leaf Segmentation According to the Verticality
In order to duplicate the number of images and allow our algorithm to recognize
the different parts of a wheat leaf, we proceeded to a second segmentation. It is a
separation step of the three leaf parts according to the verticality. Figure 4 shows the
obtained images. We got three new images. The first representing the top of the leaf,
the second the middle and the third the bottom of the leaf.
Resize and Data Augmentation
We have resized the wheat images to 400 * 300 format because the high resolu-
tion of the images requires powerful equipment for training. Large image sizes
bring to memory problems and can stop training. Image resizing was done using
the OpenCV graphics library and a resizing algorithm [15]. In order to generalize the
wheat diseases recognition on other images with different orientations and positions
of the leaves, we adopted an augmentation step. The purpose of the augmentation is
Fig. 4 Segmentation result according to the leaf verticality: a top, b middle, c bottom
and provides better prediction accuracy. Indeed, when a model has learned too much
the peculiarities of training data, it has a very high success rate (up to 100%). Instead
of setting generalized learning on the data test, the model projects closely training
results which provides an unstable model. There are no generic or statistical means
to determine the best hyper-parameters. It’s based on experimentation with empirical
values. The hyper-parameters that we study in this document are:
• The size of image filter
• The Number of Epochs
• The size of the images
3 Results
Hyper-parameters are parameters whose value is defined before the start of the
learning process. They cannot be learned directly from data and must be prede-
fined. Unfortunately, there are still no statistical means to determine the best. So we
experimented with different values. We train our model and we choose at the end the
parameters that provide the best precision.
Filter Size
A filter is a matrix of values. It is generally small and often square. It is used to
show some characteristics of a given image (color, outline, brightness, sharpness,
etc.). By scanning the image in successive steps, according to the values and the size
of the filter matrix, we obtain a new image more or less modified. The purpose of
this process is to reveal some characteristics of the image. It is not best to use very
small filters as this can cause over-learning. We tested LeNet model which contains
three convolutional layers. Through tests we carried out on the size, we evaluated the
precision obtained. If a number of pixels is required to recognize the object, large
filters should be used. If the objects present a small part of the image, the small filters
are better [17]. Since the symptoms of the diseases are in a small part of the images
(for example small spots for septoria), it seems logical to use small filters to extract
the characteristics. The obtained results confirm this hypothesis (Fig. 5). By reducing
the size of the filter, we obtained a better precision, we reached a precision of 97%
with a filter size (3 × 3), 92% for (12 × 12) and 89% for (40 × 40).
We found similar results when we tested the small sizes of filters in the neural
network for the tomato diseases diagnosis.
achieved by the “matplotlib” visualization library, shows the evolution of two details
as a function of the epoch number. The first accuracy is that of the training data
and the second is the precision of the tested data. In all the graphs presented in this
document, the x-axis refers to the number of epochs and the y-axis corresponds to
the reached prediction rate.
By growing the number of epochs, the training takes more time and the precision
increases but stagnates from epoch 40. We also notice that the blue curve (training
data) exceeds the red curve (test curve) in the thirtieth epoch which means that
our model did over learning and therefore good performance on training data and
poor generalization on other data. We therefore worked with 25 epochs because this
number offers good precision and at the same time prevents us an over-learning.
Tairu Architecture
Experimentation with this architecture provided us a 90.47% prediction accuracy.
This is a slightly low diagnostic accuracy. However, despite the precision obtained,
we had an over-learning.
AlexNet Architecture
The “Alexnet” architecture achieved an accuracy of 94.04%. We preferred to test for
100 epochs to see the accuracy evolution of the precision. Precision increased until
epoch 60 where it stabilizes.
LeNet Architecture
By using the LeNet architecture, we obtained an accuracy of 97.61%. The precision
increases according to the time.
The synthesis of our experimentation with the three architectures mentioned
above, allowed us to adopt the appropriate model according to the precision of
predictions obtained (Fig. 7).
We obtained the best predictions from the LeNet model. Indeed, its architecture
has a limited depth which makes it more efficient when we have less data. In the case
of our study, we obtained a prediction accuracy of 97.61% since we have few images.
We could only use 400 images obtained after the segmentation and augmentation
operations. The results also show the effectiveness of CNN architectures regardless
of the data masses. The good precision of certain architectures requires thousands
of images in order to adapt to their depths (in the case of Tairu and AlexNet archi-
tectures). Others provide more relevant results with average mass of images (case
of LeNet architecture). Although it is true, it is not always easy to get millions of
images and to process them on ordinary machines or servers.
4 Discussion
for the deep learning model. The addition of other parameters could improve the
accuracy such as the location, the plant age, etc.
5 Conclusion
In this work, we have integrated deep learning for the early detection of certain
wheat diseases that spread in Morocco. The process requires RGB images of sick
and sick leaves. We have concluded from the results of our experiment that the
use of convolutional neural networks is very promising. We managed to predict
the various diseases of wheat at 97.62%. Particular attention is considered in the
availability of training data, the adequate choice of network structures and adopted
hyper-parameters.
References
1. Blancard D (2012) Tomato diseases: identification, biology and control. Manson Publishing,
London, 688 p. eBook ISBN 9780429158445
2. USDA Foreign Agricultural Service. Morocco: grain and feed update (2016). https://gain.fas.
usda.gov/Recent%20GAIN%20Publications/Grain%20and%20Feed%20Annual_Rabat_Mor
occo_3-30-2016.pdf
3. Lyamani A (1989) Fréquence et sévérité des maladies d’orge, blé dur et blé tendre dans les
principales régions céréalières du Maroc. Rapport d’activité 1988–89. INRA/Centre régional
Settat, Maroc, pp 75–76
4. Eyal Z, Sharen AL, Prescott JM, Van Ginkel M (1987) The Septoria diseases of wheat: concepts
and methods of disease management. CIMMYT, Mexico, 52 p
5. Brahimi M, Boukhalfa K, Moussaoui A (2017) Deep learning for tomato diseases: classification
and symptoms visualization. Appl Artif Intell 31(4):1–17
6. Al Hiary H, Bani SA, Reyalat M, Braik M, ALRahamneh Z (2011) Fast and accurate detection
and classification of plant diseases. Int J Comput Appl 17:31–38
7. Zubiolo A (2017) Extraction de caractéristiques et apprentissage statistique pour l’imagerie
biomédicale cellulaire et tissulaire. PhD thesis, p 134
8. Tzu-Liang T, ReddyAletia H, Zhong H, Yongji K (2015) E-quality control: a support vector
machines approach. J Comput Des Eng 2015(2):91–101
9. Padmavath J (2012) Logistic regression in feature selection in data mining. University SRM,
India, pp 1–3
10. Yang X, Guo T (2017) Machine learning in plant disease research. Eur J BioMed Res 3:6–9.
https://doi.org/10.18088/ejbmr.3.1.2016m
11. Kamilaris A, Francesc X, Prenafeta B (2018) Deep learning in agriculture: a survey. Comput
Electron Agric 147:70–90
12. Ferentinos KP (2018) Deep learning models for plant disease detection and diagnosis. Comput
Electron Agric 145:311–318
13. Lu J, Hu J, Zhao G, Mei F, Zhang C (2017) An in-field automatic wheat disease diagnosis
system. Comput Electron Agric 142:369–379
14. Karkare P (2019) First Steps — Deep Learning using Python and Keras. https://medium.com/
x8-the-ai-community/first-steps-deep-learning-using-python-and-keras-3ae6bcb37d0e
Early Detection of Wheat Diseases in Morocco … 35
15. Wang SY, Wang O, Zhang R, Owens A, Efros AA (2020) CNN-generated images are surpris-
ingly easy to spot... for now. In: Proceedings of the IEEE conference on computer vision and
pattern recognition, vol 7
16. Tairu O (2018) plant disease detection using convolutional neural network. https://towardsda
tascience.com/plant-ai-plant-disease-detection-using-convolutional-neural-network-9b58a9
6f2289
17. Oliveira DCD (2016) How do we choose the filters for the convolutional layer of a Convolu-
tion Neural Network (CNN)? https://www.researchgate.net/post/How_do_we_choose_the_fil
ters_for_the_convolutional_layer_of_a_Convolution_Neural_Network_CNN
18. Sharma A (2018) What is an epoch in deep learning? https://www.quora.com/What-is-an-
epoch-in-deep-learning
Genetic Algorithm Design Optimization
for Non-standard Spur Gears
Abstract Spur gear design is one of the understood subjects that can not be treated
by limited classical methods. For this reason, an optimization process by the genetic
algorithm was conducted in order to minimize the structure volume. Mechan-
ical problems such as gear designs present many variables to analyze and several
constraints to respect. In order to gain in terms of weight cost and transmission,
Genetic Algorithms are used to improve the design process of the gearbox and find
the optimum gear parameters. In this paper, the influence of the profile shift factor
on one spur gear volume and clearance volume is studied including the bottom clear-
ance volume equation in the fitness function. The optimal results of the model with
corrected pinion and gear are compared with the standard spur gear model without a
profile shift factor in bottom clearance equation volume or structure equation volume.
1 Introduction
Genetic algorithms are frequently used for optimization purposes to solve theoretical
understanding problems. They have been often used in different fields of science,
engineering, or biomechanics [1]. Genetic algorithms adopt the approach of organic
evolution, inspired by populations of individuals with a high quality of genetic mate-
rials. Collecting and modelling different organic evolutions lead to form the basis of
GA [2]. The greatest application of this technique is in the optimization sector, as
discussed by De Jong [3].
Thomas Bäck, Ulrich Hammel, and Hans-Paul Schwefel talked about the history
of evolutionary computation, their objectives, and different structures, including
genetic algorithms [4].
In the mechanical field, the gearbox is known for its primary role in power trans-
mission for machines. They are often used in rough environments leading to severe
damages, system failures, and high maintenance costs [5]. For this reason, mechan-
ical designers are always recommended to improve designs to give better results and
reduce costs.
Cevdet Gologlu used GA to minimize the volume of a parallel axis gearbox with
two-stage helical gear trains. He compared the design results with static and dynamic
penalty function to the solutions of a deterministic design already developed [6]. Said
Golabi and Javad Jafari Fesharaki analyzed the volume/weight minimization of one,
two, and three-stage gear train. Depending on different values of input power, the
hardness of gears and, the gear ratio, the results were compared to previous works
solutions [7].
Particle swarms (PS) are used too on many occasions in gearbox volume optimiza-
tions. Ketan Tamboli and Sunny Patel took the volume geometry for a pair of helical
gear reducer as an objective function. The results were enough to improve perfor-
mance requirements [8]. Even if sometimes the particle swarm optimization (PSO)
solutions outperform the GA results on the gearbox minimization volume/weight in
some points, Genetic algorithms still the most used method in gears optimization
design [9].
One of the most interesting parameters to study in the gears optimization field
is the profile shift factor and his influence on different aspects of design, stress,
and transmission capacity. Diez-Ibarbia studied the impact of profile shift factor on
the efficiency of spur gears. He concluded that increasing the profile shift factor
reduces the transmission efficiency [10]. Coming to Genetic Algorithm optimization
Paridhi Rai and Aman Agrawal included the profile shift as a design variable, with
the module, the face width, and the number of teeth in their volume optimization
of the helical gear, using Real Coded Genetic Algorithm (RCGA). They found that
optimum volume with profile shift factor is lower than a volume without it [11].
Daniel Miler and Antonio Loncar treated the impact of profile shift factor on
spur gear pair optimization using the Genetic Algorithm. The pair gear volume was
considered as an objective function. Its a function of addendum diameter instead of
the pitch diameter to highlight the impact of shift coefficient in the objective function.
The obtained results were confirmed by the specialized commercial software results.
The inclusion of the profile shift as a variable leads to a lower volume and weight
[12]. The same authors have made a multi-objective GA optimization on spur gears
based on volume and efficiency as objective functions. The aim was to decrease the
power loss and design volume. The optimum results were achieved by reducing the
module, the face width, and raising the profile shift and teeth number [13].
A lot of work has been achieved on gearbox design. In this work, the authors
studied the impact of the shift coefficient on one pair of spur gear volume by the
process of genetic algorithms.
Two models of the geometry structure are compared. The first model with a simple
volume equation is calculated as a function of pitch diameter, the face width b and,
the transmission ratio i. The second model included the profile shift factor in the
bottom clearance equation to make clear his influence on the optimization results.
Genetic Algorithm Design Optimization … 39
The pinion and gear shift coefficients x 1 and x 2 are taken as function variables in this
model besides the module m, the face width b, and the pinion teeth number Z 1 in the
first and second model.
The novelty of this paper is the inclusion of the bottom clearance volume equation
in the structure volume equation for more accurate results and to spotlight more the
influence of the shift coefficient on optimization results.
2 Problem Definition
The aim of this paper is the volume minimization of one simple single-stage spur
gear system transmission, then analyze the effect of the bottom clearance volume
equation, and profile shift factor of the pinion and the gear on the optimization result.
When the reference line of the basic rack is at a distance (r + y) from the spur
gear center, and conjugate with the gear involute part of the profile teeth. Then the
spur gear became a corrected gear (Non-standard), with a profile shift y [14]. The
clearance is the difference between the addendum of one gear and the addendum of
the mating gear [15].
In the solid-structure of spur gear volume computation, the tooth part is the most
complicated part to calculate. That’s because the tooth part begins from the root
circle to the pitch circle. The bottom clearance is created when the bottom of the first
gear tooth and the tip of the second operation gear tooth don’t match (Fig. 1). Sa is
the tooth tip thickness of the mating spur gear (Eq. (3)). To simplify the calculations
Sa is considered as thickness or broad of the clearance cube. Z is the number of gear
teeth, and b is the gear face width.
The clearance volume V c is described by the Eq. (2), with r is the pitch radius,
r a is the addendum radius, r b is the base radius, α is the pressure angle, and S is the
thickness of the teeth at the pitch circle.
40 S. Belarhzal and E. M. Boudi
Fig. 1 The bottom clearance between two spur gears teeth in contact
Vc = Sa ∗ h ∗ b ∗ Z (2)
ra
Sa = S ∗ − 2ra ∗ (invαa − invα) (3)
r
The relation between the profile shift of gear and the thickness of its teeth is
presented in the Eq. (4):
π
S=m∗ + 2xtanα (4)
2
Z Z Z
ra = m ∗ + x + 1 ; r = m ∗ ; r = m ∗ ∗ cosα (5)
2 2 2
rb
invαa = tanαa − αa ; αa = ar ccos (6)
ra
h is the height of the bottom clearance. It presents the radial distance between the
top of the first spur gear and the bottom of the second gear in contact. For corrected
spur gears the height of the bottom clearance is written as:
Using the bottom clearance volume while taking the profile shift factor in
consideration, transform the equation of gear volume into equation right bellow:
V = f (m, b, Z 1 ) = π b (m Z )2 − d B2 /4 − Vc (8)
3 Method
This optimization research combine between genetic algorithms method and ISO
6336:2006 Standards for spur gear strength calculations. The author’s first objective
Genetic Algorithm Design Optimization … 41
was to study the effect of the profile shift factor on structure volume. For this reason,
besides modules, pinion number of teeth, and teeth face width used in the first model,
pinion profile shift factor and wheel profile shift factor were included as functions
variable for the second model, using genetic algorithms. The optimization results for
the first and the second model were compared.
Corrected gears (with positive profile shift factor) use an extended part of the
involute to form the tooth flank compared to standard gears; (without profile shift
factor), which leads to the creation of the clearance between tooth profiles in contact
while meshing. In this paper, the authors analyze the influence of positive values
of shift coefficient on the clearance between teeth in contact. The bottom clearance
equation was considered as a fitness function besides the structure volume equation
for a multi-objective optimization using genetic algorithms. The aim is to consider
the variation of the entire volume structure as well as the bottom clearance volume
according to profile shift modifications.
The same input power, input speed, and shaft diameter are used for the three
models. The module, the face width, the number of pinion teeth, and the profile shift
factor for the gear are limited, to have possible results in practical situations. Data
about boundary conditions are presented in Table 1. The profile shift factor for the
pinion is related to the tip tooth thickness by the equation:
The objective of this paper is the minimization of the spur gear volume, which is
considered as an objective function for the genetic algorithm. To come up with the
results, the genetic algorithm toolbox of software MATLAB is used [16].
The Table 2 shows the chosen input power, torque, and rotational speed. The
gear number of teeth Z 2 is calculated using the transmission ratio i. The material
considered for this structure is the standard steel with ρ = 7830 kg/m3 as density.
The pinion and gear shaft diameters are fixed on 20 mm.
In this paper, the population size adopted is 1000 individuals with a random
vector for the initial population variables instead of the uniform vectors [17]. For
more precision, the optimization was repeated five times for each volume model.
Most spur gear failures came from bending stress on the root of the teeth and surface
durability limited by the contact stress between teeth in contact while meshing [18].
For this reason, the contact stress and the bending stress are considered as constraints
42 S. Belarhzal and E. M. Boudi
Table 1 Optimization
Input variables Range
boundary conditions
Module m [mm] [2, 10]
Pinion number of teeth Z 1 [18, 30]
Face width b [14, 65]
Pinion profile shift factor x 1 [0.2, 0.7]
Gear profile shift factor x 2 [0.2, 0.4]
Table 2. GA optimization
Parameter Value
input data
Pinion shaft diameter [mm] 20
Gear shaft diameter [mm] 20
Input power [w] 5000
input speed [tr/min] 1500
Poisson’s ratio 0.3
Elasticity factor [GPa] 210
Torque [N.mm] 35000
Transmission ratio i 2.5
Maximum bending stress σ F max [MPa] 750
Maximum contact stress σ Hmax [MPa] 450
b
g3 (x) = 6 − (12)
m
b
g4 (x) = − 12 (13)
m
Genetic Algorithm Design Optimization … 43
Changing the profile shift value could increase or decrease the contact ratio CR
value [21], therefore with a greater number of CR corrections the gearing action is
disturbed by tooth gear tip interference [22]. To avoid this undesirable phenomenon
the tooth interference constraints are used (14), and (15):
17 − Z 1
g5 (x) = − x1 (14)
17
17 − Z 2
g6 (x) = − x2 (15)
17
i is the transmission ratio.
Z2
i= = 2, 5 (16)
Z1
The optimization without profile shift factor converged faster. The program running
was for more than 30 times and didn’t present any deviation. First of all, the gener-
ation number fixed on 500 generations, then reduced to 100 generations to decrease
the computation time. The computation time could be influenced too by the initial
population. Generally, using pseudo-random numbers for initial population gives
accurate results in less time [23].
The results of the genetic algorithm optimization for both equation volume of
one-stage spur gear are presented in Table 3. It showed that the smallest volume was
obtained by including the bottom clearance volume and the profile shift factor in the
spur gear equation volume. It was reduced by 48%, the volume minimization is clear.
Increasing the profile shift factor by the amount of a positive value leads to a
higher clearance volume between the gear tooth flank as shown in Fig. 2. It can be
justified by the difference in shape between standard and corrected spur gear teeth.
During the meshing process, a prolonged part of the involute is taken as a flank.
Which explains the larger space between gear and pinion in contact.
As shown in Fig. 3, the Pareto front illustrates the best optimization solutions
between global structure volume and clearance volume. As the volume increases, so
does the clearance volume.
44 S. Belarhzal and E. M. Boudi
Fig. 2 Impact of profile shift factor on the percent of clearance volume compared to the global
structure volume
Fig. 3 Pareto front of GA multi-optimization for global structure volume and clearance volume
for corrected spur gear pair
Genetic Algorithm Design Optimization … 45
5 Conclusion
Gears are the most important part in the transmission power field. They are widely
working in the industry. For this reason, many researchers worked on how to improve
their different aspects. Gears design is a complicated problem to treat. It contains
several variables and deals with different constraints. Genetic Algorithms present a
metaheuristic approach to treat complicated gearbox design that cannot be solved
with classical methods.
In this work, the effect of the profile shift factor for pinion and gear on the genetic
algorithm volume and bottom clearance volume optimization of one-stage of the
spur gear are studied. The results are significant since the low volume was obtained
using the profile shift factor on the spur gear volume equation. With a difference of
48% less than the first model using only the module m, the face width b, and the
pinion profile shift factor Z 1 as function variables.
Including the clearance volume equation helps to have more accurate optimization
volume results and avoid interference problems that appear with corrected gears since
V c increases while increasing the profile shift factor.
It is to be noted that this study has a limitation concerning the constraint functions
involved in the genetic algorithm. Spur gear’s dynamic properties were not considered
but will be developed in future studies.
References
12. Miler D, Lončar A, Žeželj D, Domitran Z (2017) Influence of profile shift on the spur gear pair
optimization. Mech Mach Theory 117:189–197
13. Miler D, Žeželj D, Lončar A, Vučković K (2018) Multi-objective spur gear pair optimization
focused on volume and efficiency. Mech Mach Theory 125:185–195
14. Michael GM (2013) Effect of change of contact ratio on contact fatigue Michael Gmariam
these. Addis Ababa University (2013)
15. Colbourne JR (1987) The Geometry of Involute Gears. Springer, New York
16. Mei W, Na J, Yang F, Shen G, Chen J (2016) The optimal design method and standardized
mathematical model of tooth profile modification of spur gear. Math Probl Eng 1–7. https://
doi.org/10.1155/2016/6347987
17. Maaranen H, Miettinen K, Mäkelä MM (2004) Quasi-random initial population for genetic
algorithms. Comput Math Appl 47:1885–1895
18. Samya B, Bachir A, Boudi EM, Amarir I (2019) The effect of addendum factor on contact ratio
factor and contact stress for spur gears. In: 2019 7th international renewable and sustainable
energy conference (IRSEC), Agadir, Morocco, pp 1–6. https://doi.org/10.1109/IRSEC48032.
2019.9078205
19. Samya B, Boudi EM (2020) Profile shift factor’s effect on contact stress and natural frequencies
of spur gears using the finite element method. Int Rev Model Simul (IREMOS) 13:1–12
20. Samya B, Boudi EM, Bachir A, Amadane Y (2020) Analysis of profile shift factor’s effect on
bending stress of spur gears using the finite element method. In: 2020 IEEE 6th international
conference on optimization and applications (ICOA), Beni Mellal, Morocco, pp 1–6. https://
doi.org/10.1109/ICOA49421.2020.9094486
21. Gebremariam M, Thakur A, Leake E, Tilahun D (2018) Effect of change of contact ratio on
contact fatigue stress of involute spur gears. Int J Current Eng Technol 8:719–731
22. Maiti R, Roy AK (1966) Minimum tooth difference in internal-external involute gear pair.
Mech Mach Theory 31:475–485
23. Maaranen H, Miettinen K, Penttinen A (2007) On initial populations of a genetic algorithm
for continuous optimization problems. J Glob Optim 37:405. https://doi.org/10.1007/s10898-
006-9056-6
A Robust Method for Face Classification
Based on Binary Genetic Algorithm
Combined with NSVC Classifier
1 Introduction
We have chosen to use the NSVC of the classification function [5]. Next, to
demonstrate the effectiveness of our proposed algorithm GA-NSVC, we have tested
on two real Face datasets.
Our trials on the distinctive datasets show that our strategy prevailing with regards
to lessening the quantity of highlights while keeping up great precision surpassing
99%.
The remainder of this article is coordinated as follows: The essential ideas of
GA are given in area Sect. 2. Section 3 presents the NSVC classifier. In Sect. 4 we
conduct a comprehensive experimental study and we conclude the article and give
perspectives to our work in section.
Genetic algorithm is an iterative algorithm that search for the optimum of a constant
size population [3, 6]. The constant population size leads to a phenomenon of
competition between chromosomes.
The making of another populace from the past is finished by:
Selection
The choice is a cycle by which a chromosome is replicated in the new populace
dependent on the estimations of the capacity to advance for this chromosome.
Crossover
The straightforward intersection comprises on, right off the bat picking several chro-
mosomes with a likelihood p and afterward in a subsequent advance, cutting the agent
directs in an indistinguishable arbitrary situation in the two guardians. This produces
two “head fragments” and two “tail portions”. At last, we drive two guardians’ tail
portions to get two youngsters who acquire a few attributes of their folks (Fig. 1a).
Mutation
A transformation is characterized as the in-rendition of somewhat in a chromosome
(Fig. 1b). This is comparable to haphazardly change the estimation of a boundary.
Transformations go about as commotion and forestall the development to freeze.
They permit guaranteeing worldwide (just as nearby) search, contingent upon the
weight and the quantity of pieces moved. Moreover, they guarantee numerically that
the general ideal will be accomplished.
Fitness Function
The fundamental objective of highlight choice is to utilize less highlights to get the
equivalent or better presentation [1, 7]. The inquiry’s procedure will likely discover
a component subset boosting this capacity. The hybrid and change capacities are the
primary administrators that haphazardly sway the wellness esteem.
A Robust Method for Face Classification … 49
Fitness = K N N _{accuracy}
SVM is first presented by Vapnik and partners for the issues of arrangement and
relapse can be viewed as another preparation method dependent on conventional
straight, polynomial and Spiral Premise Capacity (RBF).
In any case, the suspicion that the preparation information are indistinguishably
produced from obscure likelihood circulations may restrict the use of SVM to the
issues of regular daily existence.
To loosen up the supposition of indistinguishable circulation, the NSVC [8,
9] utilizes a bunch of vicinal centers capacities fabricated dependent on directed
grouping in the element space actuated by the portion.
Signifying a non-direct change of the info space X to a high-dimensional space
utilizing a piece work as:
: Rn → F
xi → (xi ), i = 1, . . . , l
l
k = αki z i , k = 1, 2, . . . , c (1)
i=1
F = J − TH (2)
l
(x) = yi αki K(x, xi ) + b , k = 1, 2, . . . . . . , K (4)
i=1
l
l
Mkm (x) = yi yj αki αmj K xi , yj + b , k, m = 1, 2, . . . . . . , K (5)
i=1 j=1
c
f(x) = βk yk Lk (x) + b (6)
k=1
A Robust Method for Face Classification … 51
BOSS Dataset contains 2000 images with 949 faces and 1051 non-faces. Figure 2
shows some face pictures in the preparation and test sets.
MIT-CBCL Face Dataset: is a preparation set comprising on 6,977 trimmed
pictures (2,429 countenances and 4,548 nonfaces). Figure 3 shows some face pictures
in the preparation and test sets.
The classification accuracy without dimension reduction on the two datasets is given
in Table 1:
52 M. Ngadi et al.
The NSVC can improve the quality of classification, confirming that NSVC is
performing well.
In view of the GA setup (Table 2), the accompanying outcomes were acquired. The
deliberately picked wellness work empowered the GA to limit grouping mistake
from KNN.
Figures 4 and 5 shows the development of the fitness estimation of the best chro-
mosome in every age. This curve shows that the error in the classification for the best
individual that represents the subset of features to select decreases with changes in
population at every generation.
We used the GA approach to select a subset of features for the NSVC classifier
(Fig. 6).
The characterization precision on the two subsets is given in Table 3:
Our experiments on different databases show that our method has succeeded in
reducing the number of features while maintaining a good precision which exceeds
99%.
A Robust Method for Face Classification … 53
Finally, we can confirm the effectiveness of our method to select the most relevant
features and give a better classification.
54 M. Ngadi et al.
5 Conclusion
References
3. Amine A, El Akadi A, Rziza M (2009) Aboutajdine D (2009) GA-SVM and mutual information
based frequency feature selection for face recognition. INFOCOMP J Comput Sci 8(1):20–29
4. Gunavathi C, Premalatha K (2014) Performance analysis of genetic algorithm with KNN and
SVM for feature selection in tumor classification. Int J Comput Electr Autom Control Inf Eng
8(8):1490–1497
5. Yang X, Cao A, Song Q, Schaefer G, Su Y (2014) Vicinal support vector classifier using
supervised kernel-based clustering. Artif Intell Med 60(3):189–196
6. Huang C-L, Wang C-J (2006) A GA-based feature selection and parameters optimization for
support vector machines. Expert Syst Appl 31:231–240
7. Mathworks T (2013) Statistics toolbox user’s guide. The MathWorks, Inc, Natick
8. Ngadi M, Amine A, Hachimi H, El-Attar A (2016) A new optimal approach for breast cancer
diagnosis classification. Int J Imaging Robot 16(4):25–36
9. Cao A, Song Q, Yang X (2004) Mammographic mass detection by vicinal support vector
machine. In: IEEE international joint conference on neural networks, Budapest, Hungary, vol
3. IEEE, pp 1953–1958
10. Ngadi M, Amine A, Nassih B, Hachimi H, El-Attar A (2016) The performance of LBP and
NSVC combination applied to face classification. Appl Comput Intell Soft Comput 2016:10.
Article ID 8272796
11. Ngadi M, Amine A, Nassih B, Hachimi H, El-Attar A (2017) Uniformed two local binary
pattern combined with neighboring support vector classifier for classification. Int J Artif Intell
15(2):102–115
12. Ngadi M, Amine A, Hachimi H, El-Attar A (2017) Vicinal support vector classifier: a novel
approach for robust classification based on SKDA. Pattern Recognit Image Anal 27(3):403–411
13. Ngadi M, Amine A, Nassih B, Hachimi H (2019) A highly efficient system for mammographic
image classification using NSVC algorithm. Procedia Comput Sci 148:135–144
14. Nassih B, Ngadi M, Amine A, El-Attar A (2018) New proposed fusion between DCT for
feature extraction and NSVC for face classification. Cybern Inf Technol 18(2):89–97
15. Ngadi M, Amine A, Nassih B, Hachimi H, El-Attar A (2018) Intelligent classification of heart
disease using neighboring support vector classifier. Int J Tomogr Simul 31(4):54–64
A Proposed Solution to Road Traffic
Accidents Based on Fuzzy Logic Control
Halima Drissi Touzani, Sanaa Faquir, Saloua Senhaji, and Ali Yahyaouy
Abstract Road Traffic accidents are one of the main problems causing death and
fatal injuries in all countries. Several general suggestions were presented to the
ministry of transport to try to overcome this problem. This paper discusses a specific
approach based on fuzzy logic control to train semi-autonomous cars to prevent
accidents from occurring. These types of cars equipped with many devices along
with decision making algorithms based on learning will make cars more intelligent
to take the right decision when no reaction of the driver is noticed.
1 Introduction
National Highway Traffic Safety Administration (NHTSA), states that more than
90% of the crashes are caused by human errors, and 40% of deadly ones imply drunk,
drugged, tired or distracted drivers [1, 2]. Therefore the technology of Autonomous
vehicles has the possibility to impact vehicle safety, traffic jam and travel behavior
[2], so to avoid potential collision and vehicle accidents lot of studies has been applied
on cars system to became automatic or semi-automatic, Triveni et al. equipped the
car with an ultrasonic sensor which will continuously track for any obstacles from
the front side, In this way they can ensure that a safe distance is always maintained
between the two cars and thus accident can be avoided [3].
Keeping up with evolution, these last years’ vehicles are being more equipped
with AEB (Automatic Emergency Braking) systems as stated in Erik et al. article,
they have also described one of the latest AEB systems called Collision Warning with
Full Auto Brake and Pedestrian Detection (CWAB-PD) which helps drivers avoid
both rear-end and pedestrian accidents by providing a warning and, if necessary,
automatic braking using full braking power [4].
To find solutions to traffic accidents, data should be collected and analyzed to
discover insights from it and present solutions to the government. In the previous
paper, a research study on road accidents in the country of Morocco was conducted
using a real database provided by the Ministry of transport and equipment’s and
composed of 1953 Instances of data on road accidents taken in 2014. The data
has been maintained using data transformation techniques such as filters and then
analyzed by three different methods, MS Excel, K-means classifier. The two methods
showed almost similar results presenting reasonable causes on different types of
accidents which can help to provide valuable suggestions to the Ministry to reduce
road accidents in the country like for example [5]:
• Install roundabouts that help reduce collisions and control speed.
• Build overpasses and underpasses in the places knowing more traffic.
Since the above mentioned suggestions are general and mostly related to infras-
tructure, a second solution is proposed which relies on using the data gathered to train
Semi-autonomous cars to be able to take decisions. These types of cars are equipped
with many devices along with decision making algorithms based on learning such
as deep learning or linguistic variables such as fuzzy logic will make cars more
intelligent to take the right decision when no reaction of the driver is noticed.
Fuzzy logic system have the potential to improve the performance of a system
by applying reasoning and are able to describe complex systems in linguistic terms
rather than numerical values [6]. This approach is very suitable in several areas within
many researches as it represents data by rules in place of using specific equations
[7].
Fuzzy logic has been excelled in various domain such as medicine, Jason et al.
describe how this method works by representing its application in a simplified model
of fluid resuscitation of intensive care unit patients [8]. Blood pressure has been
used for regulation of depth of anesthesia by fuzzy logic control and in some cases
exceeded a human operator [9]. In energy domain, Faquir et al. adopted Fuzzy logic
to control the energy between production and consumption and also confirm the
availability of power on demand for an isolated house in the city of Essaouira in
Morocco [10]. In electric vehicle Rachid et al. apply fuzzy control to ensure a longer
life for a hybrid electrical vehicle by controlling the flow of energy inside [11].
We propose in this paper an approach based on Fuzzy logic control to help semi
autonomous cars to decide and take action when it is needed.
A Proposed Solution to Road Traffic Accidents … 59
We based our work on the previous result which analyzed 1953 instances using
statistics graphs and k-means classifier to find similarities between data and extract
different types of accidents that frequently happen and important shock points [5].
The most important similarity criteria determined by K-means algorithm are
described in Table 1. Based on the previous statistical analysis it was found that
the majority of accidents occur during the day, so we decided to partition the results
by day and night, we noticed also that the k-means algorithm could identify the most
important attributes that represents the 10 different classes and one of the attributes
was day/night. Other attributes such as the type of collision, the initial shock point
and the movement of the vehicles in the road were also considered as criteria to sepa-
rate different classes between them, so the majority of accidents that have occurred
in the day have a type of collision by the side of the car, front and by entering the
road. On the other hand, the accidents that happened at night took place between
cars of the same sense [5].
• Rules: Fuzzy Inference Rules are the most important operation of the fuzzy infer-
ence system, including a number of inference rules connecting the input fuzzy
variables to the output fuzzy variables [7], and the experts offered all the rules
and the IF-THEN conditions to control the decision making system [13].
• Fuzzification: where the fuzzifier performs the act of fuzzification, it transform
a numeric value to a fuzzy set [14], and for further processing the crisp inputs
are measured by sensors and passed into the control system. For example Human
body temperature, pressure, etc. [13].
• Inference Engine: in this step we could determine the degree of match between
fuzzy input and the rules. Based on the % match, it helps to determine which rules
need implement according to the given input field. And then the applied rules are
combined to develop the control actions [13].
• Defuzzification: which is an important operation in the process of fuzzy set,
because it transforms a fuzzy set to numeric value [14].
4 Semi-autonomous Car
A semi-autonomous car is a vehicle that dispenses with the action of the driver
in many areas. It is equipped with several sensors and software that automatically
cooperate using artificial intelligence. It will use various techniques to sense its
surrounding, such as radar, Lidar, and camera.
Lidar (Light Sensing and Telemetry), It is a technology which is based on remote
sensing that calculates distance by illuminating a target with a light beam and analyzes
the reflected light. It is mounted on the roof of the vehicle on a cylindrical enclosure
which rotates 360° [15].
Lidar consists of a transmitter, a mirror and a receiver, It can also use laser,
ultraviolet and infrared light to create images of objects.
RADAR (Radio Detection and Ranging) is becoming an important automotive
technology. This equipment helps to estimate the mutual speed of the vehicle and
the object, based on electromagnetic waves. Radar technologies generally used are:
frequency-modulated continuous wave radar or FMCW, pulse Doppler radar [16].
A Proposed Solution to Road Traffic Accidents … 61
VIDEO CAMERAS: These are attached to the top of the front window, near the
rearview mirror and reproduce real-time 3D images of the road in front of you. These
are adopted to detect traffic lights and signs, unforeseen everything like pedestrians,
or animals…. [17].
A semi-autonomous car will equipped with such devices along with decision
making algorithms based on learning such as deep learning or linguistic variables
such as fuzzy logic will make cars more intelligent to take the right decisions when
no reaction of the driver is noticed.
5 Proposed Solutions
Based on the results we found by both Excel graphs and K-means classifier algorithm,
we provided the following suggestions to the ministry of transportation to help reduce
accidents:
• Widen the most frequented roads, and install warning signs that alert drivers
who exceed the speed limit, especially where there are schools, Public services,
hospitals…
• More control over non-controlling intersection and giving more priority to turning.
• Install roundabouts that help reduce collisions and control speed.
• Build overpasses and underpasses in the places knowing more traffic.
Aging causes a loss of meaning. For example, an elderly person sees a sign only
65 m away, while a 20-year-old sees it at 100 m. So, it is important for the ministry
to work on the visibility and legibility of signage [5].
Concerning the behavior of the driver it remains relative to the education received
by each individual as well as to the values acquired by each one of them.
Since all of these suggestions are general and most of them requires building a new
infrastructure, a second solution is proposed which relies on using the data gathered
to train Semi-autonomous cars that can take decisions like breaking, reducing speed,
and stopping when it is needed in case the driver didn’t take any related action
and this to limit accidents and help ensure greater road safety. This system will be
implemented by the fuzzy logic control.
Based on results presented earlier in the data analytics phase; a scenario was generated
to help a semi autonomous car take right decisions when there is a bad or no reaction
from the driver which will prevent accidents from happening.
The Flowchart presented in Fig. 2. represents the parameters which are Type of
collision, initial shock point, Movement, braking system in S (second), and depending
on the case an action will be generated, if the driver takes action its will be no
62 H. D. Touzani et al.
reaction from the semi-autonomous car otherwise the car will take action either
braking immediately, braking and attention left or right or reducing speed.
This scenario will be Implemented using fuzzy logic control since the decisions
can be based on linguistic values.
7 Conclusion
Thousands of people die in traffic crashes yearly; this involves not only loss of human
life but also property damage, this makes road safety one of the most studied subjects
nowadays. In this paper, a data analytics was done by three different tools to analyze
a big database of road traffic accidents collected by the ministry of transportation in
Morocco. The purpose of the analysis was to identify the hidden parts in the data
and locate the most important initial points and the types of intersection in the road
that mostly causes accidents. As a better solution, a semi-autonomous car equipped
A Proposed Solution to Road Traffic Accidents … 63
with different types of sensors and cameras connected to intelligent models can help
the car take the right decisions when there is no driver reaction. This paper presented
an algorithm based on fuzzy logic to consider different scenarios and try to help the
car take the right decision and prevent the accident from occurring when the driver
doesn’t react.
Our future work will be the implementation of the algorithm and try it on different
scenarios.
References
Abstract This paper presents the optimal allocation of the distributed generations
(DGs) into the distribution networks (DNs) using the Archimedes Optimization Algo-
rithm (AOA). The AOA optimizes three DGs operating at unity power factor repre-
senting the PV renewable energy sources (RES). The standard 33-bus DN is used as
a test system to verify the AOA’s effectiveness in both single-objective and multi-
objective optimizations. In addition to the overall power loss elimination, the total
voltage deviation (TVD) of the DN is reduced by the optimum allocation of three
DGs. In multi-objective optimization, the Pareto Optimal Front (POF) method is
adopted to determine the non-dominated solutions. A fuzzy linear function decides
the Best Compromise Solution (BCS) between the points set by the POF. The AOA’s
obtained results in both single and multi-objective optimizations are compared to
the Particle Swarm Optimization (PSO) and Atom Search Optimization (ASO) algo-
rithms. The three algorithms are effective in solving the optimization problem or
both single and multiple dimensions. Moreover, the AOA outperforms the ASO and
PSO algorithms in different case studies.
A. Eid (B)
Department of Electrical Engineering, Aswan University, Aswan 81542, Egypt
e-mail: [email protected]
Department of Electrical Engineering, College of Engineering, Qassim University, Unaizah
56452, Saudi Arabia
H. El-Kishky
Electrical Engineering Department, The University of Texas at Tyler, Tyler, TX 75799, USA
e-mail: [email protected]
1 Introduction
In recent years, environmental issues, fuel cost volatility, energy liberalization, and
technological developments have contributed to the rise in distributed generation
(DG) investment. This development has generated fantastic benefits but produced
a range of difficulties in planning and operating distribution networks (DNs). The
primary objective of inserting a DG into any DN is to provide extra real and/or reactive
power feeding the load that reduces the total power supplied by the mains. More
advantages to utilities of DG allocations include power loss and TVD reductions,
enhancing the stability of DNs, increasing the system reliability [1–3]. At the same
time, DG technologies provide a secure operation of the electricity markets during
the deregulation of the power industry. Integration of DG units into DNs plays a
vital role in providing support for reactive power, spinning reserve, and frequency
control. However, poorly designed and incorrectly controlled DGs will contribute to
reversing power flows, increasing power losses, and resulting in feeders’ overload.
Recently, many researchers applied different metaheuristic optimization algo-
rithms to allocate DG units into DNs. In [4, 5], the authors used the particle
swarm optimization (PSO) algorithm or enhanced versions to reduce the distribu-
tion networks’ power losses. Grey wolf optimizer (GWO) algorithm is adopted in
[6, 7] to minimize the power losses and reactive losses and enhance the voltage
profiles of different DNs. A bacterial foraging optimization algorithm (BFOA) has
been applied in [8, 9] to optimally determine the size and site of multiple DG units
with DNs. The study also included allocating STATCOM devices with different load
models for power loss minimization, voltage profile improvement, and increasing
the reliability of the systems. A hybrid of ant and artificial bee colonies is applied
in [10] for optimal placement and sizing of different DG sources to minimize the
power losses and energy costs and enhance voltage stability. An enhanced artificial
ecosystem-based optimization (EAEO) algorithm is developed and applied in [11]
for optimal allocations of multiple DG units to reduce the power losses and TVD and
enhance the system stability. The analysis included the application of the DG under
fixed, unity, and optimal power factors. The authors of [12] applied a combination
of the genetic multi-objective solver (GMOS) with linear programming to optimize
the DG units attached to DNs for reducing energy price arbitrage, energy losses and
enhance the power quality of the systems. In [13] adaptive PSO and modified grav-
itational search algorithm (GSA) are adopted to optimize the size and locations of
different DG units with single and multi-objective problems. The main objectives
were to reduce the total power losses and TVD and to enhance voltage stability.
A hybrid salp swarm algorithm (SSA) with the loss sensitivity is used in [14] for
renewable energy integration considering the annual load growth.
The authors in [15] have proposed a newly published algorithm known as
Archimedes Optimization Algorithm (AOA). It is one of the most recent nature-
inspired algorithms formulated from the physics Archimedes law of gravity. It
imitates the concept of buoyant force imposed on an entity, partly or entirely
submerged in a fluid, which is proportional to the weight of the fluid being displaced.
Multi-objective Archimedes Optimization Algorithm … 67
The application of AOA for DG allocation in distribution systems has not been
considered yet, which encourages us to implement it to deal with these optimization
problems.
In this paper, the AOA is implemented to determine the optimal sizes and locations
of three DG units operating with UPF to reduce the power losses and TVD of the 33-
bus DN. The obtained results of AOA are compared favorably with other algorithms
such as PSO and ASO. The results show the superiority of the AOA algorithm in the
single- and multi-objective optimization case studies.
2 Problem Formulation
The DG units’ integration to the DNs enhances their performance in terms of loss
reduction and improvement of the voltage profile. The AOA algorithm is imple-
mented to optimally determine the sizes and locations of three DG units to minimize
the total power losses and the total voltage deviation. The DG type includes the
PV operating at unity power factor (PF). The Forward/Backward sweep method is
used for solving the DN before and after the DG integrations. For single-objective
optimization, the branch power loss is calculated as:
br
Ploss = Ibr
2
Rbr (1)
where Ibr represents the branch current and Rbr represents the branch current.
The total power loss of the system, Ploss, is:
N br
Ploss = 3 br
Ploss (2)
1
where N br is the number of branches and three accounts for the three-phase
system, the power loss objective function is expressed as:
The TVD is the sum of the voltage differences between the bus voltages and unity,
as:
N b
TVD = |Vi − 1.0| (4)
i=1
where N b represents the number of buses, and Vi is the bus voltage, the TVD
objective function is as:
f T V D = min(T V D) (5)
For multi-objective (MO) optimization, both functions are optimized together as:
68 A. Eid and H. El-Kishky
In this case, the non-dominated points are determined using the Pareto Optimal
Front (POF) method. The fuzzy-based function selects the BCS in the POF collection
of points. For each objective function f i with a solution range of k points, the linear
membership function µik is calculated as:
⎧
⎪
⎨1 f i ≤ f imin
f imax − f i
µik = f imin ≤ f i ≤ f imax (7)
⎪
⎩
f imax − f imin
0 fi ≥ f imax
where f imin is the minimum value while f imax is the maximum value of the i th
objective function; Nob is the objective function numbers and Nnd is the number of
the non-dominated solutions. The minimum value of this set is the BCS of the POF
points.
3 Optimization Algorithms
The newly published AOA [15] is implemented here to optimize the three DG units
in order to minimize power loss and TVD. The performance of the AOA is compared
to the ASO [16] and PSO [13] algorithms. The details of each algorithm are found
in the respective literature.
In this section, the obtained results of the adopted optimization algorithms are demon-
strated and compared. The study includes the sizing and siting of two different RES
of PVs and WTs in the 33-bus RDS to decrease the total power losses and total
voltage deviation. Three RESs are optimized, working at unity power factor (UPF).
The following case studies are investigated.
Multi-objective Archimedes Optimization Algorithm … 69
23 24 25 26 27 28 29 30 31 32 33
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
19 20 21 22
In this paper, the optimal placement and sizing of three PVs and three WTs are
proposed using the AOA algorithm to reduce the losses and TVD. The study includes
both single-objective optimization (SOO) and multi-objective optimization (MOO)
based on a fuzzy function to determine the BCS among the POF points. In all case
studies, the obtained results of the AOA are compared to ASO and PSO algorithms.
In this section, the single-objective functions of power loss and TVD are optimized
separately using the AOA, ASO, and PSO algorithms. The obtained results are
compared in terms of total power loss, reactive loss, and TVD. The comparison
also includes statistical analysis showing the average, median, standard deviation,
and standard error for each algorithm.
Power Loss Reduction: The three optimization algorithms PSO, ASO, and AOA,
are simulated and compared, as listed in Table 1 with the base case for minimum loss
objective function. Investigating the table shows that the AOA algorithm obtains the
lowest power loss of 72.7837 kW while it is a little bigger of 72.9759 kW by using the
70 A. Eid and H. El-Kishky
Table 1 Optimal sizing of three UPF DGs with power loss reduction as the objective function
Parameters No DGs PSO ASO AOA
Power loss (kW) 210.9822 72.7951 72.8291 72.7837
Reactive loss (kVar) 143.0313 50.6530 50.7958 50.6530
TVD 1.8044 0.6164 0.6185 0.6164
Minimum voltage (pu) 0.9038 0.9686 0.9683 0.9687
DG size & location – 801.80/13 742.03/14 801.79/13
1091.31/24 1081.62/24 1091.31/24
1053.60/30 1114.29/30 1053.59/30
ASO algorithm and 72.7951 kW for the PSO case. The power loss for the base case
is 210.9822 kW. Allocating three DGs working at UPF of optimal sizes, as shown
in Table 1 at optimal buses 13, 24, and 30, reduces the power loss by 65.5% and
64.6% for reactive power loss. Moreover, TVD is reduced by 65.8%, which means
improving the system’s voltage profile. As a result, the minimum voltage of the base
case of 0.9038 pu has been elevated to 0.9687 pu after DGs allocations. The complete
voltage profile is shown in Fig. 2. The node voltages are improved at all buses after
the DG allocations. Although the three algorithms’ power loss values are relatively
equal, the AOA outperforms the other algorithms regarding the convergence speed
of the objective function against the iteration number, as shown in Fig. 3. The PSO
and ASO algorithms come in the second and third stages, respectively, compared to
the AOA algorithm.
Statistical analysis is performed to compare the performance of the three algo-
rithms for solving the DG allocation problem. Every algorithm solves the system 30
consecutive times, allocating the three DGs to minimize the power loss. The average,
median, standard deviation, and standard error are calculated for each algorithm and
listed in Table 2. It is clear from the table that AOA is the best algorithm in all listed
parameters. These results show the superiority of the AOA algorithm compared to
others.
Reduction of TVD: In the same manner as the previous case of power loss reduction,
the three algorithms are simulated to reduce the TVD as a single-objective function.
The TVD values are 0.0715, 0.0693, and 0.0636 V, respectively, for the PSO, ASO,
and AOA, as listed in Table 3. The AOA outperforms the other two algorithms of
ASO and PSO. The allocation of three DGs also contributes to decrease the power
losses and enhance the voltage profile, as shown in Fig. 4. The bus voltages are
all improved (except the slack bus) and approximately equal to unity, where the
minimum bus voltage is 0.9943 pu. The speed convergence of the three algorithms
is shown in Fig. 5. It is clear that the AOA is much faster and more accurate than the
other PSO and ASO algorithms. A statistical analysis based on 30 consecutive runs
of the three algorithms to minimize the TVD as presented in Table 4. The statistical
parameters include average, mean, median, standard deviation, and standard error.
Multi-objective Archimedes Optimization Algorithm … 71
Fig. 2 Voltage profile before and after DG allocations with power loss reduction
This section presents the multi-objective (MO) optimization of both the power losses
and the T V D of the distribution system. The PSO, ASO, and AOA algorithms are
applied to the 33-bus system to optimize the two objective functions. The decision-
maker decides the best operation point depending on the system situations and
72 A. Eid and H. El-Kishky
Table 3 Optimal sizing of three UPF DGs with TVD reduction as the objective function
Parameters No DGs PSO ASO AOA
TVD 1.8044 0.0715 0.0693 0.0636
Power loss (kW) 210.98 162.18 114.88 113.18
Reactive loss (kVar) 143.03 114.91 79.89 78.95
Minimum voltage (pu) 0.9038 0.9637 0.9944 0.9943
DG size/location – 1350.6/12 1170.0/13 1168.9/13
1481.7/24 1762.3/24 1558.7/24
1759.6/30 1782.1/30 1847.9/30
0.96
0.94
0.92
0.9
0.88
0 2 4 6 8 10 12 14 16 18 20 22 24 26 28 30 32 34
Bus number
Fig. 4 Voltage profile before and after DG allocations with TVD reduction
attributed conditions. The power loss (Ploss) and T V D objectives are optimized by
optimally sizing and sitting different DG units.
The MO-AOA, MO-ASO, and MO-PSO algorithms are used to optimize a single
DG operating at UPF. The objective is to minimize both the power losses and the
T V D using POF techniques accompanied by the fuzzy distribution function. The
POF set of points and the BCS points for all the algorithms are shown in Fig. 6.
Although the BCS points for the three algorithms are close, the MO-AOA has the
minimum values for Ploss and T V D. With two DGs, the optimal solutions with their
BCS points are shown in Fig. 6. For three DGs optimization, the POF points with the
Multi-objective Archimedes Optimization Algorithm … 73
Fig. 6 Optimal solutions for optimizing a single-DG (left) and two-DG (right)
BCS points of the different algorithms are shown in Fig. 7. In this case, the MO-ASO
and MO-PSO solutions are close together regarding the BCS points. The MO-AOA
has a lower BCS for both objectives, and hence, it beats the other algorithms. If one
investigates Figs. 6, 7 carefully, it is clear that the expansion ranges of the AOA
when optimizing the TVD and power losses are larger than with the PSO and ASO
algorithms. Subsequently, the AOA finds more non-dominated solutions than the
other algorithms. Moreover, the effectiveness of the AOA appears more clearly with
increasing the number of DG units. The BCS is much lower for AOA than for other
algorithms with three DG units, as shown in Fig. 7. Hence, the AOA is more suitable
for large DNs operations.
74 A. Eid and H. El-Kishky
5 Conclusions
The optimal sizing and siting of multiple distributed generations are determined
using the new metaheuristic Archimedes optimization algorithm (AOA) to reduce
total power losses and total voltage deviation. The AOA is implemented in a single-
objective as well as multi-objective optimizations. It is found that the allocation of
the DG units to the distribution networks decreases the power losses and enhances
the voltage profile of the system. The AOA’s obtained results surpass those obtained
by the ASO or PSO algorithms regarding the optimized functions or convergence
speed.
References
8. Devabalaji KR, Ravi K (2016) Optimal size and siting of multiple DG and DSTATCOM in
radial distribution system using bacterial foraging optimization algorithm. Ain Shams Eng J
7(3):959–971
9. Mohammadi M, Rozbahani AM, Montazeri M (2016) Multi criteria simultaneous planning
of passive filters and distributed generation simultaneously in distribution system considering
nonlinear loads with adaptive bacterial foraging optimization approach. Int J Electr Power
Energy Syst 79:253–262
10. Kefayat M, Lashkar Ara A, Nabavi Niaki SA (2015) A hybrid of ant colony optimization and
artificial bee colony algorithm for probabilistic optimal placement and sizing of distributed
energy resources. Energy Convers Manag 92:149–161
11. Eid A, Kamel S, Korashy A, Khurshaid T (2020) an enhanced artificial ecosystem-based
optimization for optimal allocation of multiple distributed generations. IEEE Access 8:178493–
178513
12. Jannesar MR, Sedighi A, Savaghebi M, Anvari-Moghadam A, Guerrero JM (2020) Optimal
multi-objective integration of photovoltaic, wind turbine, and battery energy storage in
distribution networks. J Energy Manag Technol 4(4):76–83
13. Eid A (2020) Allocation of distributed generations in radial distribution systems using adaptive
PSO and modified GSA multi-objective optimizations. Alexandria Eng J 59(6):4771–4786
14. Abdel-Mawgoud H, Kamel S, Yu J, Jurado F (2019) Hybrid Salp Swarm Algorithm for inte-
grating renewable distributed energy resources in distribution systems considering annual load
growth. J King Saud Univ Comput Inf Sci (2019, in press)
15. Hashim FA, Hussain K, Houssein EH, Mabrouk MS, Al-Atabany W (2020) Archimedes opti-
mization algorithm: a new metaheuristic algorithm for solving optimization problems. Appl
Intell 51(3):1531–1551
16. Zhao W, Wang L, Zhang Z (2019) Atom search optimization and its application to solve a
hydrogeologic parameter estimation problem. Knowl-Based Syst 163:283–304
17. Aman MM, Jasmon GB, Bakar AHA, Mokhlis H (2014) A new approach for optimum simul-
taneous multi-DG distributed generation units placement and sizing based on maximization
of system loadability using HPSO (hybrid particle swarm optimization) algorithm. Energy
66:202–215
The Recurrent Neural Network
for Program Synthesis
Abstract The program synthesis is a new and ambitious field consists of generating
from short and natural descriptions a source code. The program synthesis is rapidly
becoming a popular research problem, interesting potential has been offered by using
artificial intelligence to support new tools in almost all areas of program analysis
and software engineering. The goal of our long-term research project is to enable
and facilitate any user to create an application from a description written in natural
language by specifying the need for a complete system. This involves carrying out
a study of the user’s needs, the design and implementation of an intelligent system
allowing the automatic realization of an Informatic project (architecture, initialization
scripts, configuration, etc.) expressed in natural language. We propose a recurrent
neural network with LSTM cells and we publish a dataset of more than 145 765
questions and their best answer, especially in Java. The performance of the proposed
model is very interesting such that the success rate is 94%.
Keywords Artificial intelligence · The program synthesis · RNN · LSTM and new
dataset
1 Introduction
A. Berrajaa (B)
INSA Euro-Mediterranean UEMF, Fez, Morocco
e-mail: [email protected]
E. H. Ettifouri
Novelis Lab, Paris, France
compositional nature of the source code, and the architecture of the layers in the soft-
ware make it too difficult to generate the source code from natural language descrip-
tions. Artificial intelligence offers interesting potential to support new tools in almost
all areas of program analysis and software engineering. In recent years, artificial intel-
ligence has grown exponentially in the field of natural language processing (NLP) and
software engineering. For example, researchers have developed for several languages
several tools for analyzing and extracting useful information in digital documents
in order to generate automatic coding [2]. However, the differences between the
linguistic structures of different languages and the compositional nature of the code
do not always make it possible to extend the use of the programs developed for a
given language.
Program synthesis occupies an important place in the field of artificial intelligence,
and consists of automatically building a program from a specification [2]. It is a field
of multidisciplinary research, which makes collaborate linguists, computer scientists,
logicians, lexicographers or translators. Computing has long regarded programs as
data objects. With the rise of machine learning, there is renewed interest in treating
programs as data to feed into learning algorithms. However, programs have special
features, including several layers of structure, such as the contextless syntax structure
of a program, name and type constraints that are not without context, and program
semantics. As a result, programs do not start in a form that is immediately compatible
with most standard learning techniques [3]. Overall, the program synthesis consists
in designing a solution allowing the automatic generation of the source code of an
application based on descriptions and specifications written in natural languages
(French/English) taking into account the context and the field of activity of this
application, which forces the proposal of solutions of artificial intelligence [4].
This paper presents a recurrent neural network (RNN) with LSTM cells with a
learning process based on a new data set from the program synthesis. The new dataset
(Corpus) is based on over 145 765 Java Q&A from different topics from the “Stack
OverFlow” website.
In the rest of the paper, we start with the Sect. 2, reviewing recent work on program
synthesis, also a description of some algorithms which are considered to be the most
used and presenting models of machine learning and deep learning. In Sect. 3, we
will present the problem statement, then in Sect. 4 we present the construction of the
corpus and we describe our network of recurrent neurons with LSTM cells. In the
Sect. 5, we discuss our results and we will conclude with some thoughts on future
work.
2 Related Work
Linking a natural language description with source code has many topical applica-
tions, such as program synthesis, comment or text prediction, code autocompletion,
prediction of properties of a program, bugs detection and anomalies, the automation
of the migration and the analysis of the quality of the code. The goal of program
The Recurrent Neural Network for Program Synthesis 79
structures, recursive neural networks [16, 17] and a grammar of recurrent neural
networks [18] were used. Recursive neural networks have also shown good results
for the analysis of [19] and dependency [20] semantics. Reed and Freitas [21] and
Neelakantan et al. [22] introduce differentiable neuronal architectures for program
induction. This is an interesting emerging area of research [23], but one that does
not yet adapt to the kinds of problems that the programming language and software
engineering community is considering [24].
The generation of source code from descriptions and sentences written in natural
language (Text-to-Code) aims to help developers, end-users, to easily generate
programs. There is growing interest in this area, with applications such as converting
natural language to: Excel software macros [25], Java expressions [26], Shell
commands [27], if-then simple programs [8], regular expressions [28], SQL queries
[29], variable names, methods and their types [30, 31] and entity names [32].
This work presents a network of recurrent neurons (RNN) with LSTM cells with
a learning process based on a new dataset (Corpus) that contains more than 145
765 questions and their best response specifically in Java in order to automatically
generate source code for java applications with an n-tier (multilayer) architecture.
3 Problem Statement
new set of data more than 145 765 questions and their best answer, especially in
Java. This dataSet represents the learning data of our recurrent neural network with
LSTM.
The dataset we propose in this article is based on the dataset provided by the “Stack
Overflow” site. We rely on it because it is both the largest and the newest platform for
exchange between computer programming experts. Stack Overflow offers questions
and answers on a wide range of topics related to computer programming (java, c, c++,
hp, javascript, python, …). It is part of the Stack Exchange site network. The proposed
dataset is based on the questions and their best answers available in stackoverflow.
The main goal behind creating a new data set is to enable the scientific community to
evaluate new techniques using the same test conditions. Also, this dataset represents
an interesting and recent sample of questions and answers from computer experts
which will allow it to be a good learning sample, and it will be a strong point for a
learning model because it will be learned from experts in the field. These benefits
are usually not yet offered by the program synthesis researchers.
The dataset we use in this article is based on all the data provided by the Stack
OverFlow Questions and Answers website via the Kaggle web platform in a pre-
release version. We used this dataset because it is both an interesting and recent
sample of the questions and answers from the Stack OverFlow site. Initially, we did
the extraction of dataset containing a 1.2 m sample of questions and answers since
2008 until October 2016 [33] and to have a larger and updated sample with the new
scripts, we have added a second dataset with 143 690 questions and answers since
October 2016 until September 2020. Then we processed the data of the two datasets
checking the validity of the fields (empty fields and quotation marks) and we filtered
the lines taking only the questions and their best response from the java language,
furthermore we combined the two samples to have a single structure. The result is a
dataset of 145 765 questions and their best answer specifically in Java. For the latter
to be usable for machine learning.
82 A. Berrajaa and E. H. Ettifouri
Although there are several differences and several architectures in neural network
language models that have been successfully applied so far, all present and share
some basic principles like:
The input sequence (the input words), in our implementation it is encoded by an
encoding from 1 to K, where K is the unique number of letters in our dataSet. This
involves listing single letters and the code using the One Hot Encoding [35]. Note
that the sequenes are generated in 64 batch with 10% diversity (in the objective to
avoid overfitting).
The model makes a prediction of a sequence where the first element of each
sequence represents the next letter of the input. Concerning the network topology, an
architecture with five hidden layers has been applied each with 128 LSTM cells, or
each uses the “relu” activation function. While the output layer, a “softmax” activation
function is used to produce properly normalized probability values in order to choose
a single output letter.
As a learning criterion, Adam optimizers is used with a precision of 0,001. It
is generally advisable to normalize the input data of a neural network [36], which
means that a linear transformation is applied so that the data has zero mean and unit
variance. Only one normalization was to use is to keep all the letters in lower case.
Otherwise we kept the symbols because they represent part of the source code. The
following algorithm summarizes the main steps.
5 Experimental Results
In this section, we present the types and percentage of data in our dataSet, also the
learning rate achieved by our neural network.
6 Conclusion
In order to design and propose a model that meets the need for program synthesis.
Specifically, the generation of frontend/backend web applications based on Java tech-
nology and which respect an architecture with n levels (multilayer). We proposed
a project architecture in order to meet the different user needs expressed in natural
language on different software engineering domains. We started by introducing a
recurrent neural network with LSTM cells and we publish a set of learning data
of more than 145 765 questions and their best answer, especially in Java. The
performance of the proposed model is very interesting such that the success rate
is 94%.
The perspectives are set by proposing new methods, new neural network
architectures and implementing them in GPUs to improve the learning face.
References
1. Gulwani S, Polozov O, Singh R (2017) Program synthesis. Found Trends Program Lang 4(1–
2):1–119
2. Berrajaa A, Ettifouri EH, Dahhane W, Bouchentouf T, Rahmoun M (2019) Nl2code: a corpus
and semantic parser for natural language to code. In: International conference on smart
information and communication technologies. Springer, Cham, pp 592–599
3. Jozefowicz R, Vinyals O, Schuster M, Shazeer N, Wu Y (2016) Exploring the limits of language
modeling. arXiv preprint arXiv:1602.02410
4. Ling W, Grefenstette E, Hermann KM, Kočiský T, Senior A, Wang F, Blunsom P (2016) Latent
predictor networks for code generation. arXiv preprint arXiv:1603.06744
5. Manna Z, Waldinger RJ (1971) Toward automatic program synthesis. Commun ACM
14(3):151–165
6. Lau T (2001) Programming by demonstration: a machine learning approach. Doctoral
dissertation
7. Banarescu L, Bonial C, Cai S, Georgescu M, Griffitt K, Hermjakob U, Knight K, Koehn
P, Palmer M, Schneider N (2013) Abstract meaning representation for sembanking. In:
Proceedings of the 7th linguistic annotation workshop and interoperability with discourse,
pp 178–186
8. Quirk C, Mooney R, Galley M (2015) Language to code: learning semantic parsers for if-this-
then-that recipes. In: Proceedings of the 53rd annual meeting of the association for computa-
tional linguistics and the 7th international joint conference on natural language processing, vol
1, pp 878–888
9. Berant J, Chou A, Frostig R, Liang P (2013) Semantic parsing on freebase from question-
answer pairs. In: Proceedings of the 2013 conference on empirical methods in natural language
processing, pp 1533–1544
10. Bengio Y, Ducharme R, Vincent P, Jauvin C (2003) A neural probabilistic language model. J
Mach Learn Res 3:1137–1155
11. Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks.
In: Advances in neural information processing systems, pp 3104–3112
12. Bahdanau D, Cho K, Bengio Y (2014) Neural machine translation by jointly learning to align
and translate. arXiv preprint arXiv:1409.0473
13. Wu Y, Schuster M, Chen Z, Le QV, Norouzi M, Macherey W, Krikun M, Cao Y, Gao Q,
Macherey K, Klingner J (2016) Google’s neural machine translation system: bridging the gap
between human and machine translation. arXiv preprint arXiv:1609.08144
86 A. Berrajaa and E. H. Ettifouri
14. Chen H, Huang S, Chiang D, Chen J (2017) Improved neural machine translation with a
syntax-aware encoder and decoder. arXiv preprint arXiv:1707.05436
15. Xie P, Xing E (2017) A constituent-centric neural architecture for reading comprehension.
In: Proceedings of the 55th annual meeting of the association for computational linguistics
(Volume 1: Long Papers), vol 1, pp 1405–1414
16. Goller C, Kuchler A (1996) Learning task-dependent distributed representations by backpropa-
gation through structure. In: Proceedings of international conference on neural networks (ICNN
1996), vol 1, pp 347–352
17. Socher R, Lin CC, Manning C, Ng AY (2011) Parsing natural scenes and natural language with
recursive neural networks. In: Proceedings of the 28th international conference on machine
learning (ICML 2011), pp 129–136
18. Dyer C, Kuncoro A, Ballesteros M, Smith NA (2016) Recurrent neural network grammars.
arXiv preprint arXiv:1602.07776
19. Tai KS, Socher R, Manning CD (2015) Improved semantic representations from treestructured
long short-term memory networks. arXiv preprint arXiv:1503.00075
20. Zhu C, Qiu X, Chen X, Huang X (2015) A re-ranking model for dependency parser with
recursive convolutional neural network. arXiv preprint arXiv:1505.05667
21. Scott R, de Freitas N (2016) Neural programmer-interpreters. In: Proceedings of the
international conference on learning representations (ICLR)
22. Neelakantan A, Le QV, Sutskever I (2015) Neural programmer: inducing latent programs with
gradient descent. arXiv preprint arXiv:1511.04834
23. Kant N (2018) Recent advances in neural program synthesis. arXiv preprint arXiv:1802.02353
24. Gaunt AL, Brockschmidt M, Singh R, Kushman N, Kohli P, Taylor J, Tarlow D (2016) Terpret:
a probabilistic programming language for program induction. arXiv preprint arXiv:1608.04428
25. Gulwani S, Marron M (2014) Nlyze: interactive programming by natural language for spread-
sheet data analysis and manipulation. In: Proceedings of the 2014 ACM SIGMOD international
conference on management of data. ACM, pp 803–814
26. Gvero T, Kuncak V (2015) Synthesizing java expressions from free-form queries. Acm Sigplan
Notices 50(10):416–432
27. Lin XV, Wang C, Zettlemoyer L, Ernst MD (2018) Nl2bash: a corpus and semantic parser for
natural language interface to the linux operating system. arXiv preprint arXiv:1802.08979
28. Kushman N, Barzilay R (2013) Using semantic unification to generate regular expressions
from natural language. In: Proceedings of the 2013 conference of the North American chapter
of the association for computational linguistics: human language technologies, pp 826–836
29. Zhong V, Xiong C, Socher R (2017) Seq2sql: generating structured queries from natural
language using reinforcement learning. arXiv preprint arXiv:1709.00103
30. Alon U, Zilberstein M, Levy O, Yahav E (2019) code2vec: learning distributed representations
of code. Proc ACM Program Lang 3(POPL):40
31. Gao S, Chen C, Xing Z, Ma Y, Song W, Lin SW (2019) A neural model for method name gener-
ation from functional description. In: 2019 IEEE 26th international conference on software
analysis, evolution and reengineering (SANER). IEEE, pp 414–421
32. Parvez MR, Chakraborty S, Ray B, Chang KW (2018) Building language models for text with
named entities. arXiv preprint arXiv:1805.04836
33. StackOverflow (2008–2016) Stacksample: 10 percent of stack overflow questions and answers.
https://www.kaggle.com/stackoverflow/stacksample
34. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
35. Buckman J, Roy A, Raffel C, Goodfellow I (2018) Thermometer encoding: one hot way to
resist adversarial examples
36. Bishop CM (1995) Neural networks for pattern recognition. Oxford University Press, Oxford
MaskNet: CNN for Real-Time Face Mask
Detection Based on Deep Learning
Techniques
1 Introduction
The World Health Organization and several countries have recommended that masks
can limit the spread of certain respiratory viral diseases, including COVID-19 [1].
This recommendation is meant to reduce the spread of the disease by asymptomatic
and pre-symptomatic individuals and is a complementary measure to established
preventive measures such as social distancing. In this paper we have proposed a
solution for mask detection that can be very effectively used at airports to detect
travelers without masks and at other public places like Malls, schools, etc. Figure 1
presents the general architecture of face mask detection, in three phases: face detec-
tion, feature extraction, and classification. From the existing method of face recog-
nition, we have used the Viola-Jones algorithm [2, 3] to detect and crop faces from
images. We get the image with the face and run it through the Viola-Jones algorithm.
If the face is found, the classifier will give the region of interest of the face, and pass
it to the proposed CNN. CNN with Feature Detection Layers are used to extract the
deep features and those features are passed to Classification Layers. Softmax layer
in CNN classifies the images as With Mask or Without Mask in real time with the
probabilities of each classes.
The remainder of this paper is organized as follows. Section 2 introduces a brief
description of the related work. The Sect. 3 describes the proposed solution and
approach adopted to prepare the developed CNN models. The obtained results are
discussed in Sect. 4. Finally, in Sect. 5, conclusions, as well as suggestions for future
research, are presented.
2 Related Work
In this section, a review of the previous methods and approaches for classification
images based on deep learning and computer vision techniques will be provided.
Ququ Chen and Lei Sang [4] proposed Face mask recognition to prevent fraud and
to limit the problem of identifying the face and mask in the field of financial security
precaution. For that, OpenCV [5] and Dlib [6] are used to recognize the face and
extract it, and they can predict whether the image of which they test is a human face
or a mask. Loey et al. [7] propose a hybrid model using deep and classical machine
learning for face mask detection using Resnet50. In this study, we detect the region
of interest for helping proposed algorithms to learn more precise features for the
classification. For that, we use the cascade object detector uses the Viola-Jones algo-
rithm for detecting and extracting face from images, the images extracted will act
as the datasets for training and testing the Convolutional Neural Networks CNNs
proposed. The Viola-Jones object detection algorithm [2, 3] is the first object detec-
tion framework used in real-time, which was proposed by Paul Viola and Michael
MaskNet: CNN for Real-Time Face Mask Detection … 89
Fig. 2 With mask and without mask image samples from Prajna Bhandary dataset
Jones in 2001. It can be used to detect various object classes, but it was mainly used
for the problem of face detection.
3 Proposed Method
This section provides an overview of the proposed method: Dataset description and
approach adopted to prepare the developed CNN models that will be able to classify
whether the person is wearing a mask or not.
This study will focus on the Prajna Bhandary dataset [8]. It contains the full compo-
nent of the dataset for training and validation the proposed CNN model. For dataset
creation, Prajna takes normal images of faces and adds face masks to them using
custom computer vision Python script (Fig. 2).
In the rest of this section, an overview of the proposed approach (see Fig. 3) to prepare
a convolutional neural network model will be provided.
The proposed approach consists of five main steps:
90 I. Nasri et al.
• Step 1: Load and Explore images: Loading images from Prajna Bhandary dataset
for processing.
• Step 2: Detecting and Cropping faces from images: In the second step, we use the
cascade object detector that uses the Viola-Jones algorithm to detect and crop faces
from images (see Fig. 8). These images will be used to train and test the proposed
models. In the present study, datasets utilized are explained in two categories as
training and test sets, 70% for training and 30% for validation and testing.
• Step 3: Creating and Configuring Network Layers: In this step, we define the
convolutional neural networks CNN [9] architecture MaskNet.
• Step 4: Model training and testing: The cropped face from dataset will provided
as the input to the training algorithm, detailed in (algorithm 1).
• Step 5: Extract the model: Finally, the CNN model can be saved as a file and used
to classify images with the predicted label and probability. Figure 4 illustrates the
flowchart of the proposed algorithm for mask detection.
For transfer learning, four pre-trained network such as convolutional neural networks
CNN namely AlexNet [10], VGG16 [11], GoogLeNet [12] and ResNet18 [13] were
used to classify images as with mask or without mask. Table 1 shows an overview
of the pretrained CNN used in this work. AlexNet is a CNN that contains eight
layers; 3 fully connected layers and 5 convolutional layers. The AlexNet can classify
images into 1000 object classes, such as a laptop, pen and a variety of objects. This
network was trained on ImageNet database, with inputs at size 227 × 227 × 3 images.
VGG16 consists of 16 layers; 13 convolutional layers and 3 fully connected layers;
the convolutional layers are followed by the rectified linear unit (ReLU) layer for
faster and more effective training, this layer performs a threshold operation to each
element of the input by mapping any negative values to zero and maintaining positive
values. The image input size of VGG16 is 224 × 224 × 3. GoogLeNet is a CNN
that that contains 22 layers, each convolutional layer followed by pooling layer for
simplifies the output by reducing the number of parameters that the network needs
to learn about.
MaskNet: CNN for Real-Time Face Mask Detection … 91
Figure 5 shows the training progress plot for transfer learning using AlexNet,
and Fig. 6 shows the classification result with the predicted label and the predicted
probability.
92 I. Nasri et al.
Fig. 6 Classification images with the predicted label and the predicted probability %
For training from scratch, we creating and configuring network layers by defining the
convolutional neural network architecture. In the proposed CNN model MaskNet,
we use 3 convolutional layers and one fully connected layer. Softmax classifier is
used to classify images as With Mask or Without Mask. The network was trained by
the following algorithm. Figure 7 shows the architecture of the proposed deep CNN
MaskNet.
MaskNet: CNN for Real-Time Face Mask Detection … 93
4 Experimental Results
The processor for training and test processing platform was a 3.6 GHz Intel (R)
Core (TM) Core i5-8350U with 8 GB memory and 256 GB SSD hard disk, and
the development platform for the algorithm was Matlab R2018b. About the dataset
processing, the faces was detected and cropped from images (see Fig. 8) using the
Viola-Jones algorithm as it appears in the Table 2 below.
Once the face mask detector is trained from scratch based on our proposed CNN
model, we can then proceed on to loading our CNN mask detector, performing face
detection first, and then classify each face as with or without mask. For testing the
accuracy of the proposed model we use a dataset including images and videos of
subjects of various ethnicities, scenarios, and conditions.
94 I. Nasri et al.
Table 3 Comparison between the proposed models (CNN Transfer & CNN Scratch)
Model Accuracy Computer time Size Nbr images
AlexNet 96% 12 min 4 s 621Mbit 1325
MaskNet 99% 15 min 14 s 2.12Mbit 1325
MaskNet: CNN for Real-Time Face Mask Detection … 95
100
80
Accuracy % 60
40
20
0
VGG16 ResNet18 GoogLeNet AlexNet MaskNet
100 91 97 80 75 95
500 93 98 91 89 98
1325 96 98 97 96 99
30 1325
25
20
15
10
5
0
AlexNet MaskNet VGG16 GoogLeNet ResNet18
96 I. Nasri et al.
Accuracy %
Accuracy %
80% 80%
20% 20%
0% 0%
0.0001 0.001 0.01 0.1 0.0001 0.001 0.01 0.1
5 Conclusion
In this work, we have proposed a method for face mask detection using deep learning
techniques. The face is detected using the Viola-Jones algorithm. The proposed
convolutional neural network with Feature Detection Layers is used to extract the
deep features, and those features are passed to Classification Layers. A Softmax
layer in the CNN provides the classification output as With Mask or Without Mask
with the probabilities for each class. The proposed model has been trained and eval-
uated using the Prajna Bhandary dataset by two commonly used approaches for deep
learning: transfer learning and training from scratch. For that, we have used four
pre-trained networks, and one proposed CNN, namely MaskNet. The performances
of the models such as accuracy and computer time have been evaluated by modifying
the number of images dataset and the learning Rate value. Further work will focus
on the implementation of the proposed CNN MaskNet in an embedded system for
mask detection in the era of the COVID-19 pandemic.
MaskNet: CNN for Real-Time Face Mask Detection … 97
References
1. World Health Organization. Coronavirus disease (COVID-19) advice for the public. https://
www.who.int/emergencies/diseases/novel-coronavirus-2019/advice-for-public. Accessed 20
Oct 2020
2. Viola P, Jones M (2001) Rapid object detection using a boosted cascade of simple features.
In: The 2001 IEEE computer society conference on computer vision and pattern recognition,
CVPR 2001, vol 1. IEEE
3. Viola P, Jones M (2001) Robust real-time object detection. Int J Comput Vis 4(34–47):4
4. Chen Q, Sang L (2018) Face-mask recognition for fraud prevention using Gaussian mixture
model. J Vis Commun Image Represent 55:795–801
5. Bradski G, Kaehler A (2008) Learning OpenCV: computer vision with the OpenCV library.
O’Reilly Media, Inc.
6. Dlib C++ toolkit, 08 January 2018. https://dlib.net/
7. Loey M, Manogaran G, Taha MHN, Khalifa NEM (2020) A hybrid deep transfer learning
model with machine learning methods for face mask detection in the era of the COVID-19
pandemic. Measurement 167:108288
8. Prajna Bhandary dataset. Mask classifier. https://github.com/prajnasb/observations/tree/mas
ter/experiements/data
9. O’Shea K, Nash R (2015) An introduction to convolutional neural networks. arXiv pre-print
arXiv:1511.08458
10. Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional
neural networks. In: Advances in neural information processing systems
11. Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image
recognition. arXiv preprint arXiv:1409.1556
12. Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich
A (2015) Going deeper with convolutions. In: Proceedings of the IEEE conference on computer
vision and pattern recognition, pp 1–9
13. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In:
Proceedings of the IEEE conference on computer vision and pattern recognition
Machine Learning System for Fraud
Detection. A Methodological Approach
for a Development Platform
Abstract The democratization and massification use of credit cards lead inexorably
to a high number of fraudulent transactions. Generally, the fraud detection is part
of the anomaly detection problem. In this field, current approaches and techniques
are constantly looking for optimized solutions to detect anomalies. Faced with a
massive and growing data volume, these methods are put to the test, and thus lead
to a large number of undetected anomalies. Real time fraud detection requires the
design and implementation of scalable techniques capable of ingesting and analyzing
massive amounts of data continuously. Recent advances in storage, data analytics
processing, and open-source solutions open up new perspectives in the anomaly
detection field and in particular fraud. In this article, we are interested in the design of
a fraud detection system (FDS) based on open-sources Big Data technologies. Thus, a
general methodology is proposed based on the formalization, the implementation and
the technical design of a platform for fraud detection. The formalization part consists
of four layers: distributed storage, data processing, model building, and finally the
model evaluation. The implementation part uses Spark distributed data processing
system. In particular, we are based on its framework dedicated to machine learning,
called MLlib. The technical design part of the platform is based on the latest Big
Data technologies such as Hadoop, Yarn, Livy etc.
1 Introduction
In recent years, the volume of credit card transactions has increased considerably
due to the popularization of their uses, the rapid development of associated services,
such as e-commerce, e-finance and all mobile payments. The large-scale adoption of
credit card coupled with the development of different use cases where transactions
take place without rigid verification and supervision inevitably result in large losses
[1]. In this area, it becomes necessary, even crucial, to immediately process the
collected data to detect any potential fraud. However, traditional detection tools
are unable to handle the captured data volume [2]. Therefore, they cannot detect
anomalies and threats. It has therefore become essential to overcome the bottlenecks
of existing techniques by using a new generation of artificial intelligence approaches,
such as those based on machine learning. These approaches must work closely with
techniques for data processing, known as Big Data. They must also provide near
real-time responses.
Ultimately, we believe that Machine Learning (ML) models can provide a reliable
solution to the anomaly detection problem and therefore its application to the fraud
detection [3]. Depending on the needs and scope of ML, there are several scenarios,
how these models can be constructed and applied. Recent work in the field of ML
highlights two kinds of important platforms: development platform and deployment
platform.
Development platforms dramatically reduce the time and cost of building ML
models. They guarantee the organization of all levels of maturity of these models.
They can be out of the box or tailor-made, based on open-sources or commercial
software. Some development platforms can act as model deployment platforms as
well, and in a few cases, model life cycle management platforms, but their core is
running model building pipelines. The deployment platform brings the models to
the production environment. Suppose an ML model has been created and evaluated
on a given development platform, the next step is to use that model in a production
application. The latter can be a simple web application, embedded in an IoT device,
or a service or micro-service forming part of a complex architecture. This is the role
of the deployment platform.
The two platforms are linked by what is known as the machine learning pipeline.
Indeed, the life cycle of a model should not stop at the development and evaluation
stage. It is put into production and then returned to the development stage as much
as possible to be refined and improved.
In this work, we present a methodology for building ML models in a development
platform approach based on open-source software. This platform dedicated to a fraud
detection system (FDS) brings two important contributions:
– A fraud detection methodology which contains most of the design ideas common
in the latest FDS, which greatly facilitates the integration of detection algorithms
into the workflow.
– A four-layer framework that includes a distributed storage layer, data processing
layer, model building layer, and finally model evaluation layer.
Machine Learning System for Fraud Detection … 101
– A development platform whose central core is based on Hadoop and Spark. With
these technologies, we are able to build a scalable and reliable system.
The remainder of the paper is organized as follows. In Sect. 2, data and system
challenges for fraud detection are described. Section 3 describes our methodology
for building ML models for fraud detection. The results of the experimental evalua-
tion are illustrated in Sect. 4. Section 5 gives all the details about the development
platform. Finally, conclusion and future work terminate our article.
The design of an effective, real-time, and scalable based FDS is subjected to several
data and system challenges enumerated as follows [4–7]:
– Data imbalance: generally, fraudulent transactions represent less than 0.05\% of
total transactions. This ratio results in the very imbalanced dataset. A good FDS
should be able to handle asymmetric distributions of data.
– Data overlap: some fraudulent patterns behave like normal ones. A good FDS
should be able to detect fraudulent transactions that mimic legitimate transactions.
– Concept Drift: FDS targeting fraudulent behavior suffer from the fact that in the
real world, the profile of normal and fraudulent behavior changes over time. So, a
model that performs well over a period of time can quickly become obsolete and
produces inaccurate predictions. Therefore, a good FDS should be dynamic and
able to adapt to changes in fraudulent patterns.
– Evaluation metric: evaluation metric used for fraud detection techniques should
be chosen with care. Indeed, some measures such as accuracy are not suitable for
asymmetric distribution.
– Misclassification cost problem: cost of misclassifying each transaction varies.
Fraud detection is a very cost sensitive area. Generally, undetected fraudulent
transactions are much more serious and costly than the detection of normal
behavior as fraud. Therefore, a good FDS should prioritize transactions with
a higher misclassification cost.
– Lack of Real-Time FDS: most of the existing FDS reported in the literature work
on archival data to drive future security policies. The fraud must be detected and
blocked-in real time to avoid future fraud.
The fraud detection is a Big Data problem. Our ML methodology and application
to fraud detection takes into account all the data properties intrinsic to a Big Data
102 S. El Hajjami et al.
system. It includes four main layers: data collection and storage, data processing,
model building and evaluation before pre-production.
Model development begins by partitioning the datasets into one dataset used to train
a model, another dataset used to test the trained model. This splitting of the data
ensures that the model does not remember a particular subset of data. There are two
steps in a learning model building:
Machine Learning System for Fraud Detection … 103
– Training phase: mainly consists of the building or adjusting a model. The funda-
mental goal is to learn a pattern of trends that lend themselves well to generaliza-
tion to new data instead of simply memorizing the data he was able to see during
his training.
– Test phase: once the model is trained, it is important to check whether it behaves
correctly on new examples that are not used for training the model. To do this, the
model is used to predict the response on the test dataset. Then, the predicted target
is compared to the actual response to measure the performance of the model.
It’s important to test, measure and monitor the performance of a predictive model. We
must then define measures to be used for performance evaluation. The assessment
metrics used depends on several factors. Such as, the task of modeling (classification,
regression or segmentation), the context of the problem we are trying to solve as
well as the data distribution. These metrics are also used to compare the model’s
performance and select those that give the best performance.
For data collection, we use the Kaggle Credit Card Fraud Detection dataset1 to
illustrate these different steps. It contains the transactions carried out by credit card
during two days of September 2013 by holders of European cards. The Table 1
provides statistics for the dataset and shows that the minority class (fraud) accounts
for 0.172% of all transactions. Therefore, this dataset is very imbalanced [8]. It
contains 31 digital features. Because some of the input features contain financial
information, the PCA transformation of 28 digital input features (named V 1 , …,
V 28 ) was performed due to privacy concerns. Three of the given features have not
been transformed. Time feature displays the time between the first transaction and
each other transaction in the dataset. The Amount feature is the value of the amount
1 Anonymized credit card transactions labeled as fraudulent or genuin From Kaggle https://www.
kaggle.com/mlg-ulb/creditcardfraud.
104 S. El Hajjami et al.
spent in a single credit card transaction. The Class feature represents labels and takes
two values: value 1 for fraudulent transaction and value 0 otherwise.
During the processing phase, we proceed to remove missing data. Then, it was
analyzed and all features except Amount and Time were scaled using the PCA
transformation technique. Therefore, the Time and Amount columns are scaled and
normalized to ensure consistency.
For selection features, we are only interested in data features that are able to
separate the two classes (fraudulent, normal). Visualization techniques can be helpful
in this process. Consider the example shown in Fig. 1 showing the class distribution
for some features of our dataset. We can see for V 12 a significant divergence in the
distribution of the two classes. It is therefore a feature with high predictive power. We
therefore keep it when building models. Similarly, we can see for feature V 13 that the
distribution of normal transactions (majority class) corresponds to the distribution of
fraudulent transactions (minority class). This feature cannot effectively contribute to
the separation between the two classes, we eliminate it from the dataset. We carried
out this process for all the 28 features. As a result, 11 relevant features were selected
for our experiments: V 3 , V 4 , V 9 , V 10 , V 11 , V 12 , V 14 , V 16 , V 17 , V 18 and V 19 .
The main difficulties are the skewness and data overlap in fraud detection
cases. Our goal is to correctly classify the fraudulent transactions. So, we use
One Side Behavioral Noise Reduction (OSBNR) [9, 10] as a sampling approach.
This approach manages behavioral noise in order to improve the classification of
fraudulent transactions.
Since there is no rule of thumb for dividing a data set into training and test sets,
we choose the 70/30 rule for training/test sets. Several studies reported Random
Forest and Multilayer Perceptron to get the best performance [1, 3, 11, 12], this is
one of the reasons we are adopting Random Forest and Multilayer Perceptron in our
experiments:
– Random forest (RF) is an algorithm that consist of many decision trees. This
algorithm works best when there are more trees in the forest. Each decision tree
in the forest gives results. These results are merged in order to obtain a more
precise and stable prediction [13].
– MultiLayer perceptron (MLP) is an artificial neural network with direct action
which is made up of at least 3 layers of nodes: entry layer, hidden layer and exit
layer. Each node uses an activation function. The activation function calculates
the weighted sum of its inputs and adds a bias. This allows us to decide which
neuron should be removed and not taken into account in the external connections.
Various metrics can be used to measure the predictive accuracy of a model. In our
case, the major challenge is to tackle the imbalance problem, since legitimate trans-
actions are much more numerous than fraudulent transactions (less than 1% of total
transactions). This problem often leads to extremely high accuracy where a model
can reach up to 99% of the prediction accuracy, ignoring the 1% of minority class
cases. In other words, accuracy does not reflect reality in this data imbalance case.
The ML model performance is evaluated using AUC (Area Under The Curve)
of the ROC Curve (Receiver Operating Characteristics) [14]. The ROC curve is
generated by plotting the true positives rate TPR, against the false positives rate
FPR, on all decision thresholds. The true positive rate TPR is the proportion of
real positives (fraud) that are correctly identified as the positive class, and the false
positive rate FPR measures the proportion of real positives (fraud) that are wrongly
classified. AUC (1) is a concise measure of the performance of the ROC curve with
a single value between 0 and 1, where a perfect model has a score close to 1. In
addition, AUC has been shown to be effective for class imbalance [14, 15].
0.96
0.94
0.92
AUC Metric
0.9
0.88
0.86
0.84
0.82
0.8
MLP MLP_CNN MLP_ENN MLP_AKNN MLP_RENN MLP_TL MLP_OSS MLP_NCR MLP_OSBNR
ML Models used
Case (60%,40%)
0.96
0.94
0.9
0.88
0.86
0.84
0.82
0.8
RF RF_CNN RF_ENN RF_AKNN RF_RENN RF_TL RF_OSS RF_NCR RF_OSBNR
ML Models used
Figures 2 and 3 show some results analysis done for both MLP anf RF models
using the AUC metric. They also show the model’s performance of a new approach
called OSBNR. In~\cite{Salma2020b,el2020machine}, MLP and RF algorithms are
combined with different resampling methods to study the performance of the resulting
ML models.
Horizontal scaling ensures the platform profitability when data and processing
requirements vary between models. Adding more machines to a cluster would not
be of much value on its own. What we need is a system that can take advantage of
horizontal scalability and that runs on multiple machines seamlessly, regardless of
the number of machines in the cluster. The choice of a distributed system for the ML
model development platform is then necessary because it operates transparently on
a cluster of machines and automatically manages the necessary resources. Figure 4
shows our cluster formed by a master node and three slave nodes. The cluster is
Machine Learning System for Fraud Detection … 107
managed by a private cloud-type platform within the L3i laboratory as part of the
FEDER-PLAIBDE project2 . It is based on:
– Operating system: Ubuntu-16.04.3-server-amd64
– Hard disk: type SSD 500 Go
– Processors: 16 Go de RAM (4 sockets, 4 cores)
– Network: 2x10 Go SFP +
In this design, we made the choice of data storage in raw format. Thus, the system
does not depend on any data model. It is therefore natural that the choice of Apache
Hadoop Distributed File System (HDFS) is necessary to manage our distributed
data system. HDFS is a well-known example of a distributed file system in the big
data ecosystem. In HDFS, we use a «master/slave» architecture consisting of a single
NameNode which manages the distributed file system and several DataNodes, which
generally reside on each node of the cluster and manage the physical disks attached
to this node as well as data that is stored physically.
2 https://plaibde.ayaline.com/
108 S. El Hajjami et al.
Acknowledgment This work is carried out thanks to the support of the European Union through
the PLAIBDE project of the FEDER-FSE operational program for the Nouvelle-Aquitaine region,
France.
References
1. Hajjami SE, Malki J, Berrada M, Bouziane F (2020) Machine learning for anomaly detection.
Performance study considering anomaly distribution in an imbalanced dataset. In: Cloudtech
2020. IEEE
2. Roh Y, Heo G, Whang SE (2018) A survey on data collection for machine learning: a big data
- AI integration perspective. arXiv:abs/1811.03402
3. Dal Pozzolo A, Caelen O, Le Borgne YA, Waterschoot S (2014) Bontempi G (2014) Learned
lessons in credit card fraud detection from a practitioner perspective. Expert Syst Appl
41(10):4915–4928
4. Maes S, Tuyls K, Vanschoenwinkel B, Manderick B (2002) Credit card fraud detection using
Bayesian and neural networks. In: Proceedings of the 1st international Naiso congress on neuro
fuzzy technologies, pp 261–270
5. Şahin YG, Duman E (2011) Detecting credit card fraud by decision trees and support vector
machines
6. Adewumi AO, Akinyelu AA (2017) A survey of machine-learning and nature-inspired based
credit card fraud detection techniques. Int J Syst Assur Eng Manage 8(2):937–953
7. Puh M, Brkić L (2019) Detecting credit card fraud using selected machine learning algorithms.
In: 42nd international convention on information and communication technology, electronics
and microelectronics. IEEE, pp 1250–1255
8. Dal Pozzolo A, Caelen O, Johnson RA, Bontempi G (2015) Calibrating probability with
undersampling for unbalanced classification. In: IEEE symposium series on computational
intelligence. IEEE, pp 159–166
9. Hajjami SE, Malki J, Bouju A, Berrada M (2020) Machine learning facing behavioral noise
problem in an imbalanced data using one side behavioral noise reduction: application to a fraud
detection. J Comput Inf Eng
10. Hajjami SE, Malki J, Bouju A, Berrada M (2020) A machine learning based approach to
reduce behavioral noise problem in an imbalanced data: application to a fraud detection. In:
International conference on intelligent data science technologies and applications (IDSTA).
IEEE, pp 11–20
11. Bhattacharyya S, Jha S, Tharakunnel K, Westland JC (2011) Dataminingforcredit card fraud:
a comparative study. Decis Support Syst 50(3):602–613
12. Dal Pozzolo A, Boracchi G, Caelen O, Alippi C, Bontempi G (2015) Credit card fraud detec-
tion and concept-drift adaptation with delayed supervised information. In: International joint
conference on Neural networks. IEEE, pp 1–8
110 S. El Hajjami et al.
1 Introduction
The automation of work carried out by the human being in order to reduce the
execution time, quality of production, and precision has always made dreams to
researchers in the computer science field while simplifying workers’ routine activi-
ties and helping [1, 2] companies provide next-generation services. In recent years,
Artificial intelligence has progressed rapidly, approaching or even exceeding human
precision in certain cases, object detection represents one of the most significant tasks
[3] in the field of computer vision and building footprint detection and outlining satel-
lite or drone imagery is a very useful tool in many types of applications [4], from
population mapping to the monitoring of illegal development, from urban expansion
control to planning a prompt and more reliable rescue response within the case of
catastrophic events. Detecting building footprints in optical, multi-spectral satellite
or drone data is not easy to unravel in a general way due to the acute heterogeneity
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 111
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_11
112 T. Chafiq et al.
of material [5], shape, spatial and spectral patterns that will result in diverse envi-
ronmental conditions and construction activities embedded in various parts of the
world.
Deep learning, also known as artificial intelligence approaches to self-learning,
provides new doors for data processing and vision of machines [6]. The possi-
bilities of object detection and pattern recognition are currently being revolution-
ized by convolutional neural networks (CNN) that are used in remote sensing
[7]. CNN enables efficient analysis of image textures, i.e. the contextual signal of
multiple neighboring pixels, unlike common pixel-based methods. The self-learning
capabilities of CNN allow these textures to be analyzed effectively [7].
Even in complex urban scenes, accurate information on urban objects automat-
ically detected by drone images is a fundamental approach to obtaining a reliable
classification result [6]. At this stage, the technologies of remote sensing have become
a promising tool and are continuing to grow at an unparalleled pace [6, 8, 9]. Due
to modern platforms and sensors, such as very high-resolution satellite or unmanned
aerial vehicles (UAVs), the availability of optical Earth observation data revealing
both spatial details and high temporal urban patterns is increasing [10]. Using deep
learning methods, this optical feature space can be expanded further with details on
the 3D urban structure [11, 12]; Effective techniques are required to make full use
of this source of knowledge for urban mapping.
2 Related Works
Some of the previous work contributions use the images segmentation for the auto-
matic detection of buildings based on High Resolution Satellite (VHRS) images
are many [13–18]; in this section, two examples are cited: the works of Menarched
[17], whom was present the potential of SVMs to automatically extract buildings
in suburban area using Very High Resolution Satellite (VHRS) images. Then, SVM
classifier was used to extract buildings. The proposed method has been applied on a
suburban area in Tetouan city (Morocco) and 83.76% of existing buildings have been
extracted by only using color features [17]. Regarding the work of Vakalopoulou [19],
they work for the approach exploits a decomposed interconnected graphical model
formulation where registration similarity constraints are relaxed in the presence of
change detection wish gave the accuracy of 74% [19].
U-Net: Deep Learning for Extracting Building Boundary Collected … 113
The study of the harbor area of Agadir (Fig. 1), which is the fishermen’s village south
of Safi, Morocco. It is 30 km south of Safi, and 90 km north of Essaouira. The active
maritime population in AGADIR is estimated at 1000 sailors including fishing related
activities. The study area is covered by 1218 images with a surface area of 15 ha and a
maximum altitude of 10 m. Determination of the coordinates of the boundary markers
the first was carried out in relation to the Merchich (national Projected coordinate
system), which adopts the Lambert conformal conical projection.
4 Methodology
In this stage, ArcGIS Pro will be used to classify the UAV using statistical classifi-
cation methods. As defined above, deep learning is a form of machine learning that
depends on many layers of non-linear processing listed in a model for feature iden-
tification and pattern recognition [20]. For object recognition and image processing,
Deep Learning models can be combined with ArcGIS [21]. Using the Classification
Training Sample Manager tools to generate training samples of entities or objects of
interest in ArcGIS Pro, and then using the “Label Object for Deep Learning” tool to
label them. Finally converted to a format for the in-depth learning structure. These
training examples are used to train the model using the “Train Deep Learning Model”
rid analysis tool or the ArcGIS API for Python. The trained model associated with
the model definition (and possibly the Python raster function script file) is packaged
and shared as a dlpk (deep learning package) element, which is then used to generate
the raster analysis inference tools. The inference tool extracts specific features or
classifies pixels in the image. The dlpk element can be used multiple times as input
to the raster analysis tool, making it easy to evaluate multiple images at different
locations and over different time periods after model formation (Fig. 2).
The orthophoto of the Tan-Tan harbor (Fig. 3) will be used as a training dataset, then
the orthophoto’s harbor of Agadir as a test of the accuracy of the model, and produce
the vector data. In fact, the training dataset has been exported using a tool available
U-Net: Deep Learning for Extracting Building Boundary Collected … 115
Fig. 3 BF on the
Orthophoto of Tan-Tan’s
harbor
in ArcGIS Pro which is called Export Training Data For Deep Learning in order to
generate a thematic raster 8bit unsigned (Fig. 4).
The preparation of data can be a very long process involving the selection and
massaging of training in the format needed for any deep learning model [22, 23].
Typical data processing command lines include the splitting of the data into
training and test (validation) sets, various data techniques are implemented, data
structures are created to load the data into the model, batch sizing is specified, etc.
All these long-term activities can be automated by ArcGIS, and the training dataset
exported by ArcGIS is read directly by the planning process.
The first CNNs date back to the 1980s with the work on the Necognitron by K.
Fukushima [25], but it is in the 1990s that these networks became popularized with
the work of Y LeCun & al about character identification [26].
A CNN is a network of neurons whose layers are linked not by a matrix operation
but by a two-dimensional convolution. NDCs allow the direct processing of large
amounts of data such as images for several reasons:
– The images are spatially correlated (the values of adjacent pixels are generally
very close) and the convolution layers make it possible to create links between
these spatially correlated data.
– Convolution layers are usually followed by sub-sampling (pooling) layers that
greatly reduce the size of the data.
A CNN architecture is formed by a stack of independent processing layers (Fig. 5):
The complete training data was divided into training set and validation was defined
at the preparatory data level, and the validation set is by default 0,22 or 20% and the
rest of the training data is 80%. Here, it lets us know how well our models generalize
data that they have never observed, and how they avoid overfitting of the model. As
we can see in this case, our model (Fig. 7) can also categorize pixels in construction
or not with an accuracy of 96%.
118 T. Chafiq et al.
80%
60%
40%
20%
0%
1 2 3 4 5 6 7 8 9 101112131415161718192021222324252627282930
5 Results
In this step, the Geoprocessing tool which called Classify Pixels Using Deep Learning
is available in ArcGIS Pro was used to process the images and generate a classified
raster (Fig. 8).
The results of this tool will take the form of a classified raster that contains both
background and building footprints.
To extract building footprint, the Pixel Classification tool using deep learning to
segment the imagery through the model and post-processed the resulting raster in
ArcGIS Pro. For post-processing the classified imagery, the Model Builder (MB) has
been used, which is a visual programming language for building workflows [33].
U-Net: Deep Learning for Extracting Building Boundary Collected … 119
The results of the building footprint model generated on Agadir’s Harbor (Fig. 9);
this is few subsets of results overplayed on the drone recording. furthermore, it works
well when any space divides buildings from each other.
6 Conclusion
The model described in this paper offers many advantages among which the ability
to elucidate better identification and automatic detection of building-footprint based
on the Training-Model. Moreover, it shows that the benefit of using improved ground
truth labels for building segmentation without human intervention is a testament of
the progress made in AI. Furthermore, such goals have been completed by making the
system capable of recognizing buildings in different situations, achieving a precision
rate of 96% after a training period on images acquired and processed for this purpose.
This would also enable future research to focus on adding a third dimension to
generate an indexable object 3D model that could give more value and precision to
the building- footprint.
References
1. Naya F, et al (2006) Workers’ routine activity recognition using body movements and location
information. In: 2006 10th IEEE international symposium on wearable computers. IEEE
2. Chafiq T, Ouadoud M, Elboukhari K (2020) Covid-19 forecasting in Morocco using FBprophet
Facebook’s Framework in Python. Int J 9(5)
3. Ding P et al (2018) A light and faster regional convolutional neural network for object detection
in optical remote sensing images. ISPRS J Photogram Remote Sens 141:208–218
4. Martin F-M et al (2018) Using single-and multi-date UAV and satellite imagery to accurately
monitor invasive knotweed species. Remote Sens 10(10):1662
5. Darraj B, Gourja B, Faiq A, Zerraf S, Tace EM, Belaaouad S (2019) The use of concept maps
to evaluate learning strategies for nursing students. Int J Innov Technol Explor Eng 9(1)
6. Kattenborn T, Eichel J, Fassnacht FE (2019) Convolutional Neural Networks enable efficient,
accurate and fine-grained segmentation of plant species and communities from high-resolution
UAV imagery. Sci Rep 9(1):1–9
120 T. Chafiq et al.
7. Huang B, et al (2018) Large-scale semantic classification: outcome of the first year of Inria
aerial image labeling benchmark. In: IGARSS 2018–2018 IEEE international geoscience and
remote sensing symposium. IEEE
8. Corbane C et al (2015) Remote sensing for mapping natural habitats and their conservation
status–New opportunities and challenges. Int J Appl Earth Obs Geoinf 37:7–16
9. White JC et al (2016) Remote sensing technologies for enhancing forest inventories: A review.
Can J Remote Sens 42(5):619–641
10. Hirschmugl M et al (2017) Methods for mapping forest disturbance and degradation from
optical earth observation data: a review. Current Forestry Reports 3(1):32–45
11. Kattenborn T et al (2019) UAV data as alternative to field sampling to map woody invasive
species based on combined Sentinel-1 and Sentinel-2 data. Remote Sens Environ 227:61–73
12. Fassnacht FE, et al (2017) Estimating stand density, biomass and tree species from very high
resolution stereo-imagery–towards an all-in-one sensor for forestry applications? Forestry: Int
J Forest Res 90(5):613–631
13. Porway J, Wang Q, Zhu SC (2010) A hierarchical and contextual model for aerial image parsing.
Int J Comput Vision 88(2):254–283
14. Lin Y et al (2014) Rotation-invariant object detection in remote sensing images based on
radial-gradient angle. IEEE Geosci Remote Sens Lett 12(4):746–750
15. Zhang F et al (2016) Weakly supervised learning based on coupled convolutional neural
networks for aircraft detection. IEEE Trans Geosci Remote Sens 54(9):5553–5563
16. Moranduzzo T, Melgani F (2014) Detecting cars in UAV images with a catalog-based approach.
IEEE Trans Geosci Remote Sens 52(10):6356–6367
17. Benarchid O, Raissouni N (2013) Support vector machines for object based building extraction
in suburban area using very high resolution satellite images, a case study: Tetuan, Morocco.
IAES Int J Artif Intell 2(1)
18. Liu Z et al (2016) Ship rotated bounding box space for ship extraction from high-resolution
optical satellite images with complex backgrounds. IEEE Geosci Remote Sens Lett 13(8):1074–
1078
19. Vakalopoulou M, et al (2015) Simultaneous registration and change detection in multitemporal,
very high resolution remote sensing data. In: Proceedings of the IEEE conference on computer
vision and pattern recognition workshops
20. Liu W et al (2017) A survey of deep neural network architectures and their applications.
Neurocomputing 234:11–26
21. Zhang X et al (2020) How well do deep learning-based methods for land cover classification and
object detection perform on high resolution remote sensing imagery? Remote Sens 12(3):417
22. Chafiq T, et al (2018) Application of data integrity algorithm for geotechnical data quality
management12(8)
23. Ouadoud M, Chafiq T, Chkouri MY (2018) Designing an IMS-LD model for disciplinary
information space of learning management system. In: Proceedings of the 3rd international
conference on smart city applications (2018)
24. Hubel DH, Wiesel TN (1962) Receptive fields, binocular interaction and functional architecture
in the cat’s visual cortex. J Physiol 160(1):106
25. Fukushima M (1980) Dirichlet forms and Markov processes, vol. 23. Elsevier
26. LeCun Y, Denker JS, Solla SA (1990) Optimal brain damage. In: Advances in eural information
processing systems
27. Krut P (2019) MNIST Handwritten digits classification using a Convolutional
Neural Network (CNN). https://towardsdatascience.com/mnist-handwritten-digits-classific
ation-using-a-convolutional-neural-network-cnn-af5fafbc35e9. Accessed 12 Apr 2020
28. Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image
segmentation. In: International conference on Medical image computing and computer-assisted
intervention. Springer
29. Ouadoud M, et al (2019) Generate a meta-model content for collaboration space of learning
management system compatible with IMS-LD
U-Net: Deep Learning for Extracting Building Boundary Collected … 121
30. Zhou X et al (2017) Deep learning of the sectional appearances of 3D CT images for anatomical
structure segmentation based on an FCN voting method. Med Phys 44(10):5221–5233
31. Guo X, et al (2017) Deep clustering with convolutional autoencoders. In: International
conference on neural information processing. Springer
32. Forney GD Jr (1974) Convolutional codes III. Sequential decoding. Inf Control 25(3):267–297
33. Allen DW (2011) Getting to know ArcGIS modelBuilder. Esri Press
Analysis and Classification of Plant
Diseases Based on Deep Learning
1 Introduction
Plant diseases are considered to be alterations of plants that affect and may damage
their vital functions. Plant diseases are caused mainly by viruses, fungi or bacteria.
Plant diseases have a great impact on agricultural productivity. Plant diseases have
increased considerably in recent years, according to the FAO (Food and Agriculture
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 123
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_12
124 A. Ennouni et al.
Organization of the United Nations). This is mainly due to climate change, glob-
alization, and the low resilience of production systems, … Early detection of plant
diseases based on visible symptoms on leaves is one of the main challenges to protect
cultures. This mainly depends on the farmer’s knowledge and expertise. However,
not all actors in the agricultural domain are well experienced and especially small
farmers. To overcome this problem and reduce the lack of agricultural knowledge
problem, automatic detection, and recognition of plant diseases combining image
processing and artificial intelligence is a promising solution [1]. Machine Learning
(ML) and Deep Learning (DL) techniques have become more promising approaches
in the medical field because of their ability to learn from reliable and discriminating
visual features [2–4]. For their success, ML and DL have been used in many fields
including agriculture [5, 6]. Artificial Intelligence (AI) and the Internet of Things
(IOT) have been used to develop the concept of Intelligent Agriculture [7, 8].
The objective of our research is to propose a comparative study between four
different deep learning architectures for the analysis and classification of plant
diseases. Namely MobileNet, AlexNet, Inception V3, and VGG16. These archi-
tectures have been implemented and tested. Simulation results have shown that the
MobileNet architecture, even simple, has allowed better classification accuracy.
The rest of this article is as follows: in the following section, we will present the
different plant diseases and then present the different DL-based image classification
architectures and more precisely the classification of plant diseases. We will present
in detail the different DL architectures used and describe our dataset. Implementation
results will be presented afterward. And we end this work with a conclusion.
2 Plant Diseases
Plants are living beings. They are one of the three main groups in which living things
are usually classified (the other two are animals and fungi). Unlike other living beings,
most plants are fixed to the ground by roots, which make them highly dependent on
the conditions of their environment [9].
Plant diseases are damages that disrupt or modify their vital duties. All plant
species are exposed to diseases. Some plants are immune to certain kinds of diseases
while others are particularly prone to them [10]. Plant diseases can be classified as
either they are infectious or not or to the causal agent. Therefore, leaf diseases are
either fungal, bacterial, or viral [9, 10].
Bacterial plant infection symptoms are very similar to the symptoms of fungal plant
diseases. Infections are recognized by the presence of leaf spots, wilting, burns,
cankers, scabies, and rot. They can appear on fruits, leaves, stems, and/or flowers.
Analysis and Classification of Plant Diseases Based on Deep Learning 125
Fig. 1 Bacterial leaf blight on wheat and Bacterial strand test on cut stems
When they affect plants, pathogenic bacteria can be responsible for serious and
potentially disastrous diseases [11]. The following figure presents two bacterial plant
diseases (Fig. 1).
Fungi cover rust, mildew, mildew, fungus, and others are the cause of diseases that
are not risky but are harmful. There are 144000 known species. Some fungi are used
in medicine. A distinction is made between those that live in the soil or water and that
considered parasitic for plants and animals [12]. An example of the fungal disease
Sclerotinia is presented in Fig. 2.
Organisms can be infected with viruses or viroids. Viruses differ from viroids by their
protein layer. Both are mainly transmitted by insects or nematodes. The symptoms
of this infection vary and depend on the type of virus infecting. The viral disease can
affect all areas of the plant (see Fig. 3). Plant viruses are mainly isometric (polyhedral)
or stem-shaped [13].
126 A. Ennouni et al.
Computer vision has made prodigious steps in recent years thanks to machine learning
and deep learning [3, 3]. CNN (Convolutional Neural Network), denoted also as
ConvNet, is a classification algorithm that can be used as both machine learning (one
layer) and deep learning (many layers). CNN does not require much pre-processing
as other classification algorithms.
A standard CNN architecture is generally composed of a succession of convolu-
tion and pooling layers and at the end there one or more fully connected layers. An
average pooling layer replaces the fully connected in some architectures. To optimize
the performance of the CNN deep learning architecture, some regulatory units like
dropout and batch normalization are incorporated. Designed Deep Learning archi-
tectures differ essentially by the arrangement of the CNN components which play
a fundamental role in their performance. A large number of parameters have to be
defined for a given deep learning architecture such as the layers number, the neurons
number in each layer, the size of the filters mask, the weights of the neurons, the
activation function to be used, the best learning rate obtained, … [25, 25].
Several CNN architectures have been proposed in the literature. The following
table lists the most successful architectures, specifying for each the number of
parameters, the authors, the year, and the depth (Table 1).
In this work, we will conduct a comparative study between 4 Deep Learning architec-
tures for plant diseases classification. The chosen architectures are AlexNet, Incep-
tionV3, VGG16, and MobileNet. The database used contains images of diseased
crop leaves classified into 38 classes and healthy ones [23].
4.1 VGG
4.2 Inception V3
VX where X refers to the number of the version Google published. Figure 5 presents
the GoogLeNet modules.
The purpose of the start-up module is to calculate 1 × 1, 3 × 3, and 5 × 5
convolutions in the same network module. The output is then concatenated and feeds
into the next layer of the network. Compared to VGG and ResNet the Inception V3
weights (96 MB) being the smallest.
4.3 MobileNet
4.4 AlexNet
In 2012, AlexNet considerably outperformed all the prior competitors and won the
challenge by reducing the top-5 error from 26% to 15.3%. The second-place top-5
error rate, which was not a CNN variation, was around 26.2% (Fig. 7).
The network had a very similar architecture as LeNet by Yann LeCun et al. but was
deeper, with more filters per layer, and with stacked convolutional layers. It consisted
11 × 11, 5 × 5, 3 × 3, convolutions, max pooling, dropout, data augmentation, ReLU
activations, SGD with momentum. It attached ReLU activations after every convo-
lutional and fully-connected layer. AlexNet was trained for 6 days simultaneously
on two Nvidia Geforce GTX 580 GPUs which is the reason for why their network is
split into two pipelines. AlexNet was designed by the SuperVision group, consisting
of Alex Krizhevsky, Geoffrey Hinton, and Ilya Sutskever [23].
Performance Measures
In the pattern recognition field, information retrieval and automatic classification,
accuracy (or positive predictive value) are the proportion of relevant items among all
proposed items; recall (or sensitivity) is the proportion of relevant items proposed
among all relevant items. These two notions thus correspond to a design and a rele-
vance measure. Precision can thus be seen as a measure of accuracy or classification
quality, while recall is a completeness measure or classification quantity. In the multi-
class context (where the number of classes is more than 1), the recall and precision
average over all classes is calculated by the macro-mean. It calculates first the recall
and precision over each class and then calculates the average of the precision and
recall over the classes:
130 A. Ennouni et al.
n
pr ecision i
pr ecision = i=1
(1)
n
n
i=1 r ecall i
r ecall = (2)
n
n
accuracy i
r ecall = i=1
(3)
n
Where n is the number of classes and i the current class.
TPi TPi TPi + TNi
precisioni = recalli = accuracyi =
TPi + FPi TPi + FNi TPi + TNi + FPi + FNi
TP is The True Positives number, FP is the False Positives number, FP is the False
Positives number, and FN is the False Negatives number.
5 Experimental Results
Simulations have been tested on a Windows 10 computer with Intel Core i7 3.6 GHz
CPU and 16 GB of RAM.
In our study, we chose 4 different depth sizes CNN architectures: VGG16, Incep-
tionV3, AlexNet and MobileNet. The number of epoch for the 4 DL architectures is
25. Table 2 presents the evolution of training and validation accuracies, Training and
validation losses, precision, recall, and the accuracy of each of the 4 DL architectures.
As it is a multi-classification problem with 38 classes in total, we can conclude
that the 4 architectures performed well and allowed an acceptable classification
rate. The classification errors are detected particularly at the level of the “tomato
late blight" class. This is mainly due to a confusion between several classes, which
are mainly tomato related diseases, namely: “early tomato blight, septoria leaf spot
tomato blight, and tomato target spot”. Even visually, as shown in Fig. 8, the different
diseases represent visual similarities on tomato leaves. We have also noticed that
there are visual similarities between other diseases which did not allow an excellent
classification. To remedy this problem, it is possible to look for other information
that would allow the discrimination of these different diseases and to consider using
ensemble learning by concatenating features extracted from DL architectures and
handcrafted features.
The classification accuracies of Alexnet and MobileNet are better than Inception
V3. However, since the precision and recall values obtained by MobilNet are higher
than Alexnet, we can conclude that the model build using MobileNet architecture is
the most suitable to better classify plant disease.
Analysis and Classification of Plant Diseases Based on Deep Learning 131
Table 2 The evolution of Training and validation accuracies, Training and validation losses,
precision, recall, and the accuracy of each of the 3 DL architectures
Training and validaƟon Accuracy Training and ValidaƟon Loss P R A
AlexNet
6 Conclusion
Farmers usually identify plant diseases by visual assessment. The quality of the diag-
nosis depends strongly on the knowledge of the profession. However, human exper-
tise is not easily acquired by all actors in the agricultural world. Image processing
and analysis in combination with artificial intelligence represent an effective solution
to assist in the recognition and classification of plant diseases. Deep learning and
especially convolutional neural networks (CNN) was recently used in the medical
field and sees itself as a very promising solution. In this work, several DL-based
architectures such as AlexNet, VGG16, Inception V3, and MobileNet have been
tested and compared using the Plants Village database. The MobileNet architecture,
even though very simple in comparison with the others, showed very encouraging
results.
References
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 135
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_13
136 A. Jadli et al.
2 Paper Context
This section introduces the basic concepts of this paper, namely, explaining the
ERP ecosystem and features, determining our problematic boundaries, and finally
exploring related works.
2.2 Problematic
Document image classification uses different approaches and can be based on super-
vised or unsupervised feature learning, which usually requires a large amount of
labeled training data.
For example, Sherif et al. [2] presented an Exemplar-based unsupervised learning
technique to analyze document layout to improve the document image classification
efficiency without the need to expand the volume of annotated data, Jun et al. [3]
studied the use of Faster RCNN (fast regional convolution for object detection)
architecture to automatically classify documents’ internal structure, Kosaraju et al.
[4] proposed a novel structure-based convolutional neural network for document
layout analysis, called DoT-Net. DoT-Net is a multiclass classifier that can effectively
identify document regions to solve document classification tasks.
Similarly, The recent advances in Generative adversarial Networks (GAN) led
research to explore extensive use for such a concept in different domains and various
purposes [5]. Mansourifar et al. [6] introduced a new concept called Virtual Big
Data to address the problem of insufficiency of training data in many research areas,
Xu et al. [7] compared using GAN-based data augmentation for underwater image
classification with other data-augmentation techniques, Lee et al. [8] proposed an iris
image augmentation based on a conditional generative adversarial network (cGAN)
and a method for improving recognition performance that uses this augmentation
method. Meanwhile, Wang et al. [9] propose a GAN-based solution to augment
training data for improved performance of palmprint recognition.
138 A. Jadli et al.
The Generative Adversarial Network architecture and the first empirical demonstra-
tion of the approach were proposed in 2014 by Goodfellow et al. [11], describing
an architecture summarily involving a generator model that takes points from a
latent space (random noise) and generates an image, and a discriminator model that
classifies images as either real (authentic) or fake (output by the generator) (Fig. 1).
The two models are arranged in a competition (in a game theory perception)
where the generator model (Fig. 2) seeks to fool the discriminator model (Fig. 1),
and the discriminator is provided with both examples of real and generated samples
and seeks to distinguish between fake and real images.
4 Our Contribution
4.1 Overview
The experiment discussed in this paper (Fig. 3) can be divided into two different
parts:
• Using DCGAN to generate images of the scanned documents;
• construct a model for visual document classification using generated and real data.
A generator model (G) is trained on images of real-world company’s scanned
documents. The process starts from a latent vector (usually a normal distributed
random vector). It follows with the generator generating an image based on real vector
distribution. The generated image is forwarded alongside the authentic images to the
discriminator model, trying to tell the fake pictures from the real ones. The calculated
loss helps back-propagate the errors improving performance over iterations until the
discriminator cannot tell the real from the generated document images.
Therefore, the previously trained generator is used to generate a subset of fake
documents to create a new dataset, which is made of 50% of real documents and
50% of dummy document images.
A Convolutional Neural Network (CNN) algorithm is used in the next step to create
classification models out from each dataset and compare them based on models’
accuracy to assess if we can train classification models using data generated by
DCGANs.
Table 1 Distribution of
Document class Sample count
samples classes and the
number of samples E.I 180
HI 109
CH 106
RT 103
Total 498
The objective of this paper is to assess the impact of augmenting the existing
training dataset on the accuracy of the classifier in a document classification task.
The classifier used is a Convolutional Neural Network (CNN) using a simple archi-
tecture with an image input size of (64, 64) with a single channel (grayscale images).
Different techniques can be used to tackle the model’s over-fitting issue, such as max
norm constrains, L2/L1 regularization, and dropout. After comparing these methods,
dropout was simpler to use and very efficient. It was used with a rate of 0.5 despite
roughly doubling the number of iterations required to converge[10].
Despite the small size of the dataset (498 images distributed on four classes), the
generator showed good performances at generating fake document images that looked
plausibly real (Fig. 5). The generator performed in EI and HI classes better than CH
and RT classes because of the difference between the samples count in each category
(EI case) and the simplicity of standard layouts (HI case),
142 A. Jadli et al.
The second part of the experiment was using a CNN algorithm to classify the
documents on 3 distinct datasets:
• D1: contains only real authentic document images;
• D2: contains original real dataset augmented with classical data augmentation
techniques;
• D3: contains original real dataset alongside images generated in part 1 (using
DCGAN);
• D4: contains original real dataset using the conventional and DCGAN data
augmentation simultaneously.
The models built were standard CNN using the same architecture over the three
datasets. Table 2 shows the model’s performance based on accuracy.
Based on overall results, we can assess that model trained on the conventional
augmented dataset (83% accuracy) performed better than the model trained on the
original dataset (77%), and DCGAN augmented dataset (79%). Due to low vari-
ance in geometric data augmentation techniques, this can lead to model overfitting.
Combining the two approaches helped improve the model accuracy (89%), which
is a considerable performance boost even without hyper-parameter tuning or opti-
mization. Figures 6 and 7 show the confusion matrix of the model trained using the
DCGAN augmented dataset (CNN3) and the model trained using the combination
of conventional techniques and DCGAN (CNN4).
Table 2 Performance
Augmentation technique Accuracy (%)
comparison between
constructed models Raw Dataset 77.1
Classical Augmentation 83.3
DCGAN-based Augmentation 79.2
Conventional + DCGAN Augmentation 89.6
A Novel Approach to Data Augmentation … 143
This paper investigates the use of machine-generated data to substitute real data in
training machine learning algorithms. The experimental results show that despite a
small dataset size (498 documents), the use of DCGAN based data augmentation can
be a good approach in the document classification context. We can recommend this
approach to be used in practice in a lack of labeled data situation for document image
144 A. Jadli et al.
References
1. Jadli A, Hain M (2020) Automatic document classification using deep feature selection and
knowledge transfer, April 2020. https://doi.org/10.1109/IRASET48871.2020.9092256
2. Abuelwafa S, Pedersoli M, Cheriet M (2019) Unsupervised exemplar-based learning for
improved document image classification. IEEE Access 7:133738–133748. https://doi.org/10.
1109/ACCESS.2019.2940884
3. Jun C, Suhua Y, Shaofeng J (2019) Automatic classification and recognition of complex docu-
ments based on Faster RCNN. In: 2019 14th IEEE international conference on electronic
measurement and instruments, ICEMI 2019, November 2019, pp 573–577. https://doi.org/10.
1109/ICEMI46757.2019.9101847
4. Kosaraju SC, et al (2019) DoT-Net: document layout classification using texture-based CNN.
In: Proceedings of the international conference on document analysis and recognition, ICDAR,
September 2019, pp 1029–1034. https://doi.org/10.1109/ICDAR.2019.00168
5. Jadli A, Hain M (2020) Toward a deep smart waste management system based on pattern
recognition and transfer learning. In: 3rd international conference on advanced communication
technologies and networking, CommNet 2020, September 2020, pp 1–5. https://doi.org/10.
1109/CommNet49926.2020.9199615
6. Mansourifar H, Chen L, Shi W (2019) Virtual big data for GAN based data augmentation. In:
Proceedings - 2019 IEEE international conference on big data, big data 2019, December 2019,
pp 1478–1487. https://doi.org/10.1109/BigData47090.2019.9006268
7. Xu Y, Zhang Y, Wang H, Liu X (2017) Underwater image classification using deep convo-
lutional neural networks and data augmentation. In: 2017 IEEE international conference
on signal processing, communications and computing, ICSPCC 2017, December 2017, vol
2017–January, pp 1–5. https://doi.org/10.1109/ICSPCC.2017.8242527
8. Lee MB, Kim YH, Park KR (2019) Conditional generative adversarial network-based data
augmentation for enhancement of iris recognition accuracy. IEEE Access 7:122134–122152.
https://doi.org/10.1109/ACCESS.2019.2937809
9. Wang G, Kang W, Wu Q, Wang Z, Gao J (2019) Generative Adversarial Network (GAN) based
data augmentation for palmprint recognition, January 2019. https://doi.org/10.1109/DICTA.
2018.8615782
10. Antoniou A, Storkey A, Edwards H (2017) Data augmentation generative adversarial networks.
arXiv. arXiv, 12 November 2017. https://arxiv.org/abs/1711.04340v3. Accessed 17 Jan 2021
11. Goodfellow I, et al (202) Generative adversarial networks. Commun ACM 63(11): 139–144.
https://doi.org/10.1145/3422622
12. Radford A, Metz L, Chintala S (2016) Unsupervised representation learning with deep
convolutional generative adversarial networks
Knowledge Driven Paradigm
for Anomaly Detection from Tweets
Using Gated Recurrent Units
Abstract Online social media has become of great importance in the recent past.
It has proven to be a medium for connectivity as well as effortless publicity. Given
the vast scope of activities that can be performed with the help of social media, it
can also be misused. An anomaly has no concrete definition but the best description
of an anomaly would be to identify it as the one that exhibits abnormal behavior. In
this paper two events have been taken into consideration, namely black lives matter
and demonetization, that took great popularity over social media platforms in recent
years and received mixed emotions and behavioural patterns from the public. The
response to these events have been observed on twitter and an anomaly in the trend
has been predicted in the course of a selected time frame using a proposed system
model to accomplish the process efficiently. The deep learning model uses a Gated
Recurrent Unit classifier along with a domain ontology to enhance the process of
prediction. A comparative graph of other unsupervised training models have also
been included to display the efficiency accomplished through the system designed.
The overall accuracy of the trained system for demonetization is observed to be
95.43% and black lives matter is seen to be 96.18%.
1 Introduction
Technology has come to play a vital role in the connectivity of the global community.
With its speeding advancements in the discovery of new software environments to
S. Manaswini
Department of Computer Science and Engineering, SRM Institute of Science and Technology,
Ramapuram, Chennai, India
G. Deepak (B) · A. Santhanavijayan
Department of Computer Science and Engineering, National Institute of Technology,
Tiruchirappalli, India
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 145
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_14
146 S. Manaswini et al.
bind people across the globe, there are so many platforms that are available online
today, that gives people a chance to encourage thoughts, opinions, talent promotion,
news, forum for publicity of various on-going events, awareness propagation, busi-
ness and so on. To list a few of the former- Facebook, Instagram, Twitter top the charts
for maximum networking with regard to such content. Whereas anomalies are those
that exhibit irregularity in a sequence of data. They are often categorized as illegal,
sarcastic, undesired or counter-productive to the ideology. Anomalous behavior has
both positive and negative impacts. Such data can be helpful in the prediction of
natural disasters, disease out-breaks, emergency detection but on the contrary it can
also be used to promote malevolent activities such as terrorist attacks, fraudulence
using social media. It also takes into consideration, controversial standing of events
and provides a platform to express political opinions for influencing the mob and
trending topics. On a broad spectrum, social media anomalies are driven by either
single individual—‘point anomaly’ or groups of people—group or tribe anomaly
where latter generally refers to malicious behavior of hidden groups of people. These
anomalies can be described as a change in interaction patterns, change in method
of interaction and therefore can be classified into three kinds, based on the pattern
on input of the anomaly namely- content, structural and behavioral. This paper shall
be identifying two such scenarios- ‘Demonetization’ and ‘Black lives matter’ over
selective time frames. Tweets from specific time frames extracted during the peak of
both the issues have been analysed to detect anomalies in content.
Motivation: Any political action or global issue implies uncalled for public attention
wherein the entire cause of implication is twisted by individualistic opinions and the
original motive is often lost. The issues considered above impacted in ways that have
resulted in contrasting opinions. And it is interesting to determine and classify tweets
that stand against/out of topic or those that have exhibited anomalous behaviour over
the given time. The black lives matter protests have been observed over the time
frame of two months- May 20 to July 10, 2020. The mentioned time frame witnessed
a peak in the number of tweets regarding the issue. Demonetization was observed
over five months- November 10 2016 to April 20 2017.
Contribution: This paper proposes a methodology that makes use of Gated Recur-
rent Units (GRU), a kind of Recurrent Neural network (RNN) with a gating mech-
anism. A domain ontology [1, 2] derived from an RDF knowledge base has further
been compared to the former using an integration of cosine similarity and Web Point-
wise Mutual Information (Web PMI) that helps determine the prediction error or the
dissimilarity in comparison which will in turn denote the anomaly in the crawled
dataset. Twitter datasets have been specifically chosen due to the high political usage
and media attention which is crucial to the topic addressed and tweets can be conve-
niently made distinct during pre-processing for easier classification. GRU being an
improvised take on the RNN, have been observed to resolve a preliminary drawback
of the standard RNNs. GRU has resolved the dependency issue of the loss of infor-
mation in recurrent networks and has regulation over the input using structures called
‘gates’ to produce improved prediction performance over the constructed time series
model. The proposed approach has been improvised with an overall F-measure of
96.8% in black lives matter and 95.47% in demonetization.
Knowledge Driven Paradigm for Anomaly Detection … 147
Organization: The flow of the rest of the paper is as follows -: Sect. 2 consists
of related work to the field of study and experimentation. Section 3 drafts the
proposed architecture for the system of anomaly detection. Section 4 consists of
the implementation and experimental observation. Section 5 includes conclusion
and references.
2 Related Work
Sujatha et al. [3] has elaborately explained about various OSM anomalies. Behavioral
anomalies have been observed over a twitter data set on demonetization policy using
Self-Supervised learning. 86% of the public opinion was seen to support demoneti-
zation. Guo et al. [4] have put forth a Gated Recurrent Unit based Gaussian Mixture
VAE system for detection of outliers on n-dimensional time-series data called GGM-
VAE and they claim that their scheme exceeds in performance when compared to
the state-of-the-art abnormality detection model and has produced up to 5.7, 7.2%
increased accuracy accuracy and F1 score respectively, in comparison to existing
schemes. Liu et al. [5] have put together a literature review on the anomaly detection
methodologies based on social media which includes an overview of the data types
and social media data attributes, network data, temporal data anomaly detection and
also includes description of state-of-the-art technique on massive social media data.
Cho et al. [6] proposed a neural network model called RNN Encoder–Decoder that
contains a double RNN. Both were jointly trained to maximize the conditional prob-
ability of a sequence target. This paper denotes the origin of GRU. Watson et al. [7]
developed an automated outlier detection model built over supervised Long Short-
Term Memory [LSTM] and statistical analysis. Zhang et al. [8] used deep learning
methodologies to social media data for detection of traffic accidents. Deep Belief
Network (DBN) and LSTM were implemented on the extracted features and DBN
showcased an accuracy of 85%. Results displayed superiority of DBN over Support
Vector Machines (SVM) and supervised Latent Dirichlet allocation (sLDA). In [9–
20] several semantic and ontological approaches in support of the proposed work
has been discussed.
The proposed architecture as shown in Fig. 1 consists of three notable segments. The
first segment includes data pre-processing and classification using GRU. The second
segment involves the formulation of a domain ontology using an RDF knowledge
base that has been filtered out of heterogeneous fact resources. The third segment
compares the knowledge base to the tweets using an integrated cosine similarity
and Web PMI module. The very aim of this system design is to verify that the key
148 S. Manaswini et al.
indices derived from the RDF knowledge base, drawn into an ontology, is compared
to the tweets classified by GRU in order to point out an outlier in the extracted data
sequence. The first phase of the process involves scraping of the tweets through
‘Twint’, an API that facilitates scrapping of twitter.
Data pre-processing is a very vital step as the data extracted needs to be trans-
formed into a workable format. Twitter data consists of multiple symbols like ‘@’
‘#’, abbreviations, etc. that are difficult for the machine to contextually comprehend
in language processing applications. Therefore, all text was converted to lowercase
and tokenization was performed on the sequence of strings. Stemming and stop words
were removed and lemmatization is carried out to retain words but at the same time
preserve contextual meaning. On completion of the pre-processing of tweets, it is
passed into the Term Frequency- Inverse Document Frequency [TF-IDF] [21].
Prediction models are highly dependent on the feature selection process. In text
classification, features are a subset of the words in a given string. In order to classify
the ‘importance’ of a word in a document TF-IDF is used. An anomaly can only
be arbitrarily defined. In order to capture abnormal behaviour the data sequence
first needs to be in the form of a computable input in order to be analysed. TF-
IDF is computed using the product of ‘the number of occurrences of a term’ and
the ‘inverse document frequency of a word across a set of documents’. In common
terms the latter computes how less valuable a term is in the set of documents. It is
seen to approach 0 when a specific term is seen to occur frequently and otherwise
approaches 1. Equations (2) and (3) shows how Eq. (1), the final TF-IDF equation
is derived.
Nt
TF(t, d) = (2)
Nd
ND
IDF(t, D) = loge (3)
Dt
The next step is to test and classify the pre-processed tweets and dataset using
the GRU [4] classifier. The output of the classifier would be the top 20–25% of
the categorized tweet data words based on the term classes initially derived from
tweets. Further categorization, classification and comparison will be based on the top
categories. The gated recurrent neural network is an improvised RNN architecture.
It consists of two vectors or gates- reset gate and update gate. These two gates are
responsible for the regulation of information that passes on to the output gate. GRU
has been chosen as it conveniently addresses the vanishing gradient problem or the
issue of retention of necessary information in more common terms. The update gate
decides how much of the information from the preceding steps are carried onto to the
succeeding steps. The reset gate decides how much of the information is eliminated
or forgotten. The main activation functions involved in the adopted model of GRU
includes Sigmoid and tanh.
Assuming a new input ‘x’ is inserted to the network at time ‘t’, and a variable ‘y’
stores information until t-1 units, then the product of ‘x’ and its weight is summed
with the product of y and its weight and the summation is passed into the sigmoid
function which acts as the activation function and is used to contain the value between
0 and 1. Similar to the process of the update gate (1), reset gate has the exact same
procedure but the way in which the weights and gates will be used varies in terms of
function and design. When the important part of information is nearing the reset gate
is assigned a value close to 0 in order to forget the unnecessary information from the
traversal until the current point.
The current memory content is then determined by the product of the input ‘x’
and its weight that is summed with the Hadamard product of the reset gate and ‘U.y’
passed into the nonlinear tanh activation function. The final memory at time step
‘t’ is determined by the Hadamard product of the update gate and ‘y’ summed with
the product of (1−z) and the calculated current memory content ‘Ct’ and is passed
through tanh function. From (6) the final memory content at step ‘t’ is given in (7).
Equation (6) makes use of (5), Eq. (7) makes use of (66) and (4).
Current memory content C1t = tanh(x.W + Rt U.y) (6)
The second phase involves various fact sources like prevalent news daily, Google
news API, telegraph, RSS news feed and other verified resources that have been
crawled over a specified time line and tokenized to collect facts or specific terms
required for the formulation of the knowledge base. The RDF knowledge base [22]
lays on the Subject-Predicate-Object format. The source of facts is the predicate, fact
pretext is the subject and the fact context and summarized facts form the object. From
the heterogeneous sources, all the ‘keywords’ are crawled and gathered to form an
indexed fact base that leads to depiction of an in-depth domain ontology containing
all the indexed terms from the fact base. This is then compared to the categorized
tweets using the cosine similarity combined with Web PMI.
The third phase helps compare the categorized terms with the domain ontology
to help in the prediction of anomalies. The similarity function implemented is the
cosine similarity [23] along with Web PMI. Irrespective of the size of the document
cosine similarity is used to compare them. Cosine similarity is the cosine of the angle
between two vectors in a multi-dimensional space. The smaller the angle the more
the similarity.
0, H (A ∩ B) ≤ C
Web PMI (A, B) =
log2 H (A)( −NH (B)
H A∩B ) (8)
, other wise
N N
Web PMI just like every other PMI is a metric to measure the correlation between
two terms. It is observed to be directly proportional to the frequency of occurrence of
the two terms or their similarity and is defined between 0 and 1. Increasing similarity
implies that the value tends towards 1. For the purpose of anomaly prediction, the
terms from the categories classified by the GRU and those indices from the domain
ontology are compared. Because we aim on prediction of the outliers, we set the
threshold to 0.25, and any value <0.25 depicts an anomaly and >0.25 implies its
commonality in occurrence. Web PMI can be numerically defined as shown in Eq. (8),
where, A and B are the terms for comparison and N is the total number of occurrence
of synonymous term pairs and page count of the term-pair along with the page count
for individual terms is denoted by H.
4 Implementation
For the detection of anomalies the proposed algorithm has been run on a Windows
10 Operating System supported by an Intel Core i7 8th gen processor with 16 GB
RAM. The code has been trained over python version 3.9.0 using Google Collab-
oratory, which has been stacked on multiple GPUs and hosted by Jupyter note-
book. The library files included for the implementation of the proposed design
are textblob, pandas, mathplotlib, scikit learn, and NLTK. Two events have been
considered for the experiment—Black lives matter and Demonetization. Five other
Knowledge Driven Paradigm for Anomaly Detection … 151
training approaches apart from the proposed system have been adopted for obser-
vational purposes namely, self supervised learning, SVM, logistic regression, and
RNN. Precision (%), Recall (%), Accuracy (%), F-measure (%) and False Negative
Rate (FNR) has been calculated from the selected dataset.
From Fig. 2, it is evident that the proposed model has achieved an accuracy of
96.18% showing an increase of 6.04% from the adopted baseline, self supervised
learning model. The experimental precision is observed to be 98.89%, denoting
a 6.71% increase. In consideration of the baseline model [1], even though self-
supervised models have shown improved accuracy when compared to most of the
conventional unsupervised models, labelling of data is still an issue and has to be
manually annotated. Random Forest produces an accuracy of 86.43% and has a
precision of 87.15%, which clearly denotes an under-fitting model. This could be
because of the absence of discrete labels, large data sets and large memory space and
processing time that is caused by the complexity of the ensemble of the decision trees
that tests the significance of the individual variables, leading to undesired accuracy
and precision. Similarly, SVM does not perform well with large datasets and is seen
to underperform in cases that have overlapping target classes and therefore shows an
accuracy of 84.12% and a precision of 86.77%.
The Linear Regression model is seen to have an accuracy of 91.19% and a preci-
sion of 93.89%. Though there is a significant increase when compared to Random
Forest, this model assumes a linear relationship between dependent and independent
attributes that highly affects the outliers as the boundaries are tight and it might falter
in an inconsistent dataset, therefore the recall rate dipping to 86.76%. RNN shows
an accuracy rate of 91.79% but has a precision of 94.33%. One of the main prob-
lems of a standard RNN is the retention of information through the succeeding steps
and computation could be slow but with GRU the vanishing gradient problem stands
resolved and further integration with cosine similarity and Web PMI improves perfor-
mance hence proving to be the best suitable model in comparison. Similarly, for the
event demonetization as shown in Fig. 3, the accuracy measure of the proposed model
Fig. 2 Performance comparison of the proposed approach for black lives matter
152 S. Manaswini et al.
Demonezaon
Proposed Knowledge Centric Approach… 96.15 94.81 95.43 95.47
RNN 89.12 92.44 91.14 90.75
LogiscRegression 88.16 90.12 90.01 89.13
SVM 80.81 82.18 81.18 81.49
Random Forest 84.33 87.43 85.81 85.85
Self-Supervised Learning [Basepaper] 86.69 90.43 88.11 88.52
5 Conclusions
References
19. Kaushik IS, Deepak G, Santhanavijayan A (2020) QuantQueryEXP: a novel strategic approach
for query expansion based on quantum computing principles. J Discrete Math Sci Cryptography
23(2):573–584
20. Santhanavijayan A, Kumar DN, Deepak G. A novel hybridized strategy for machine translation
of indian languages. Soft Comput Signal Process 363
21. Qaiser S, Ali R (2018) Text mining: use of TF-IDF to examine the relevance of words to
documents. Int J Comput Appl (0975 – 8887) 181(1)
22. Stoermer H, Bouquet P, Palmisano I, Redavid D (2017) A context-based architecture for RDF
knowledge bases: approach, implementation and preliminary results. In: Springer international
conference on web reasoning and rule systems
23. Otsuka Y (1936) The faunal character of the Japanese Pleistocene marine Mollusca, as evidence
of the climate having become colder during the Pleistocene in Japan. Bull Biogeograph Soc
Jpn 6(16):165–170. AGCJ.
Stacked Deep Learning LSTM Model
for Daily Solar Power Time Series
Forecasting
1 Introduction
Nowadays, solar power is the most important sources for electrical power generation
[1, 2]. Solar photovoltaic produces energy from sunlight by converting solar radi-
ance into electricity [3]. The efficiency of PV installations is influenced by climate
parameters such as temperature [4]. photovoltaic energy relies on weather condi-
tions and solar irradiance, which add to photovoltaic generation heterogeneity and
instability. As stated by the IEA (International Energy Agency), The total power of
renewable electricity is projected to increase by more than 1TW during the period
starting from 2018 to 2023 [5]. Solar photovoltaic constitutes more than ½ of this
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 155
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_15
156 S. Gaizen et al.
extension and dominates the increase of renewable capacity. Taking into account the
ongoing development of solar PV sector, the electrical grid is facing the challenge of
the aforementioned heterogeneity and instability in power production. It exists many
alternatives to resolve these problems as energy storage installations to mitigate the
discontinuity of photovoltaic production [6]. Moreover, knowing how much photo-
voltaic energy would be generated, may extremely decrease power plants operating
costs and increase the resilience of the grid, thus, predicting PV power production is
a challenging task because it relies considerably on external circumstances like PV
radiation and climate conditions.
Solar power prediction strategies are primarily parted into 3 groups in the liter-
ature: physical, statistical, and machine learning configurations [7–15]. Machine
learning configurations are capable of managing complex problems [16]. It contains
3 structures: mentioned in these references [17, 18], one of these structures is the
Artificial Neural Network [18]. This structure is currently the most widely employed
strategies for time series prediction. Most traditional PV power prediction models are
confined to expose the correlation between the data, but these models are not capable
of discovering a deep correlation, uncovering hidden pattern and reveal significant
data. For the tremendous data from the recent generating stations, the application of
traditional methods isn’t conducive to ensuring reliable forecasting. Over a decade,
deep learning strategies have appeared as effective machine learning technology
[19]. They are becoming increasingly prevalent, thanks to their strong ability to
explain dependencies in time series data. Several deep learning configurations have
been recommended in these references [20–22], One of these configurations, is the
Recurrent Neural Networks [22] that takes advantage of the sequential nature of
input data. RNNs are employed to model time-dependent in timeseries data. In the
time series forecasting, RNNs provide good results, which have been proven in
several researches [23–25]. The LSTM (Long Short-Term Memory) networks is a
kind of RNNs that can accommodate information retrieval for longer time periods
[26]. Furthermore, it is largely employed for time series forecasting applications,
which is ideally tailored to problems with solar PV generation and wind power
forecasting [27]. Long short-term memory permits the memorization of the weights
that are forward and back-propagated across layers [28]. LSTM-based RNNs are an
appealing alternative for modeling consecutive data such as time series since they
integrate qualitative knowledge from past inputs. LST memory has promising oper-
ations in solar power forecasting as a novel machine learning algorithm [6, 29]. By
preserving the structure of memory cell, LSTM could sustain essential features that
should be recognized throughout the learning procedure. Thus, applying LSTM to
forecast solar power permits not only to achieve the correlation over uninterrupted
hours but also to derive its long-term behavior patterns [6]. To forecast photovoltaic
power, Abdel-Nasser [30] suggests the application of LST memory to reliably esti-
mate the generated power for solar systems. The results have exposed that the use
of LSTM provides a decrease in the forecasting error relative to other techniques.
Qing [6] suggested a new method of forecasting hourly day-to-day solar power
using LSTM-based weather forecasting. The suggested algorithm utilizes the hourly
weather as the intrant parameter, and the hourly solar power values of the similar
Stacked Deep Learning LSTM Model … 157
scheduled day as the output parameter. Due to the assumption that time dependency
is considered, the results reveal that the suggested learning algorithm is more precise
than other three models. Srivastava [29] analyzed the ability of LSTM to forecast
solar power time series, his study demonstrated the robustness of LSTM, also this
model outperforms GBR and FFNN for day-to-day solar power time series fore-
casting. Liu [31] suggested a new deep learning approach using wavelet analysis to
forecast wind speed. LST memory is applied to forecast the low frequencies of sub-
layer, and the high frequencies are forecasted by Elman-neural-network. The mixed
configuration has satisfactory performance compared to 11 traditional approaches.
Under complex weather conditions, Yuet [32] suggested an LST memory-based
approach to forecast short-term global horizontal irradiance, the study demonstrated
that LST memory surpasses ARIMA, SVR, and NN configurations, especially on
cloudy weather. Despite the mentioned LST memory techniques don’t thoroughly
analyzed experimentally the impact of various variables and structures, but these vari-
ables would influence the precision of prediction. Our article is provided to assess
the impact of the depth of the LSTM networks namely Stacked-LSTM networks.
The original time sets are separated into training and testing series. In addition, all
the outcomes of the forecasts obtained are applied to the persistence comparison
model for a proper benchmark evaluation. As a result, the Stacked-LST Memory
configuration with multiple hidden layers show high performances in the daily solar
energy forecasting.
2 Predictive Patterns
2.1 Database
The solar radiation data used in this work are from the NASA prediction of worldwide
energy resource website for the region of Fes/Morocco. The solar radiation data is
recorded within the date range from 01/01/2017 to 01/01/2019. Every 30 min, the
solar power data is registered. this recorded value was converted to daily average
data. Averaging these values would make the signal smoother, and the algorithms
are easier to learn. We observed that the forecast of solar radiation on a daily average
basis is more reliable than the prediction of solar radiation of 30 min ahead. The
data on solar radiation is standardized and the outliers and incomplete values are
excluded.
For series prediction issues, RNNs are the most widely used Neural Network archi-
tecture. It also achieved particular attention in the natural language process domain.
158 S. Gaizen et al.
The LST Memory Neural Network was suggested by [34] to prevent long-term
dependency difficulties through a targeted architecture. Unlike typical NNs, LSTM
hold a memory block to store temporary input data information.
Figure 2 illustrates the hidden layer of the LSTM, that is recognized as the LST
Memory cell.
Applying a gate structure, the LST Memory cell binds the intrant layer Xt , to the
output layer ht . There are three gates in the LST Memory cell: Intrant gate It , Output
gate Ot , and Forget gate ft . The input cell state Čt , the current cell output state Ct,
and the output state of the previous irritation, are used in the training procedure.
Using the relations given by [30], the input state, the output state and the gates can
be defined. The architecture of the LSTM can be defined as follows [30]:
Stacked Deep Learning LSTM Model … 159
Fig. 2. Architecture of
LSTM
f t = σ w f × h t−1 , xt + b f (3)
i t = σ wi × h t−1 , xt + bi (4)
ct = tanh wc × h t−1 , xt + bc (5)
ct = f t × ct−1 + i t × ct (6)
ot = σ wo × h t−1 , xt + bo (7)
h t = ot × tanh(ct ) (8)
Three error metrics were suggested to measure the efficiency of the predicting model.
The Root mean square error calculates the discrepancy between the real and the fore-
casted variables. Lower RMSE indicates better forecasting outcome, and is described
as [35]:
1
N
2
RMSE = X f − Xa (9)
N i=1
However, the Mean Absolute Percentage Error (MAPE) represents the ratio of
error to the real value [35].
1
X f − Xa
N
M AP E = (10)
N i=1 |X a |
160 S. Gaizen et al.
3 Proposed Approach
The suggested approach is designed to forecast solar power time series using deep
stacked LSTM. Figure 3 outlines the essential steps of the suggested approach.
The suggested approach comprises four primary points:
1. Collect Solar power data from the NASA prediction of universal energy resource
website for the region of Fes/Morocco. The solar power series are assembled
into training and testing period.
Data Pre-processing
Data standardization
Good Model No
yes
End
2. Pre-process and clean up the data by eliminating outliers and imputing incom-
plete values. There are several essential steps in the preprocessing of data, such
as data cleaning, data conversion, and the selection. Data cleaning and conver-
sion are techniques used to eliminate outliers and standardize data in such a way
that they take a shape that could be conveniently used to construct a model.
3. Normalize the solar power initial data.
4. Train, check and verify the Stacked-LST memory configuration. Several statis-
tical metrics are employed to measure the precision of the configuration. If the
precision is not good, the stacked LSTM layer number is increasing to assess
the performance and the depth of the stacked-LST memory model.
The future PV power is derived from the historical solar time series using the Stacked-
LST memory configuration. MAPE, RMSE and R2 are the performance assessors
for the suggested technique. In this analysis, 80% of data is applied to adjust the
configurations and the remainder is applied for assessing purposes. In addition, the
efficacy of the evolved configuration has been matched to the persistence reference
configuration. The persistence reference configuration efficiency in the daily solar
forecast was described in Fig. 4 as 1.027 for the RMSE, 27.7% for the MAPE and
0.668 for the R2 [35].
Using the stacked LST Memory configuration with four hidden layers, Fig. 5
shows the predicted solar energy values. As seen, as the depth of learning is increased,
it is apparent that a strong match between the expected and real value exists. For the
stacked LST Memory configuration, the precision of the forecast was increased by
98.7% over the persistence reference configuration.
Figure 6 demonstrates the scattering plot of the expected solar energy using the
Stacked LST Memory configuration over the real values. As shown in this figure,
the traced data points generally correlated closer towards the fitting line.
Fig. 4 Forecasted solar energy values of the Persistence Reference configuration vs. the initial
dataset [35]
162 S. Gaizen et al.
Fig. 6 Plot scattering of the expected results of the Stacked LST Memory configuration with four
hidden layers compared to the initial test data.
Table 1 displays the daily solar energy forecast results for the Stacked LST
Memory configuration. When evaluating the error values in this table, the RMSE
of 0.38, the MAPE of 10.5% and an R2 of 0.921 were found for the LSTM model
with one hidden layer. RMSE of 0.16, MAPE of 4.42% and R2 of 0.956 were LSTM
model with two hidden layers. RMSE of 0.09, MAPE of 2.486% and R2 of 0.979
were found for LSTM model with three hidden layers. RMSE of 0.016, MAPE of
0.442% and R2 of 0.997 were found for LSTM configuration with four hidden layers.
Stacked Deep Learning LSTM Model … 163
5 Conclusion
Our work provides evaluation analysis of various hybrid configuration to predict solar
energy. The precision of the forecasting is highly enhanced over to the persistence-
reference configuration. Our analysis reveals that the inclusion of hidden layers of the
LST Memory configuration improves the precision of prediction. This is illustrated
by the MAPE value beginning from 10.5% (one hidden layer) and dropping to 0.442%
when four hidden layers are included. The suggested analysis is intended to alleviate
the core problem of PV energy forecasting.
References
17. Lin G-Q, et al (2020) An improved moth-flame optimization algorithm for support vector
machine prediction of photovoltaic power generation. J Cleaner Prod 253:119966
18. Lu H, Ma X (2020) Hybrid decision tree-based machine learning models for short-term water
quality prediction. Chemosphere 249:126169
19. Hinton GE, Osindero S, Teh Y-W (2006) A fast learning algorithm for deep belief nets. Neural
Comput 18(7):1527–1554
20. Ogawa S, Mori H (2020) PV output forecasting by deep Boltzmann machines with SS-PPBSO.
Electrical Eng Jpn
21. Zang H, et al (2020) Application of functional deep belief network for estimating daily global
solar radiation: a case study in China. Energy 191:116502 (2020)
22. Jung Y, et al (2020) Long short-term memory recurrent neural network for modeling temporal
patterns in long-term power forecasting for solar PV facilities: Case study of South Korea. J
Cleaner Prod 250:119476
23. Kumar D, et al (2020) Forecasting of solar and wind power using LSTM RNN for load frequency
control in isolated microgrid. Int J Model Simul 1–13
24. Park N, Ahn HK (2019) Multi-layer RNN-based short-term photovoltaic power forecasting
using IoT dataset. In: 2019 AEIT International Annual Conference (AEIT). IEEE
25. Srivastava T, Vedanshu, Tripathi MM (2020) Predictive analysis of RNN, GBM and LSTM
network for short-term wind power forecasting. J Stat Manage Syst 23(1):33–47 (2020)
26. Zhao F, et al (2020) Time-sequenced flow field prediction in an optical spark-ignition direct-
injection engine using bidirectional recurrent neural network (bi-RNN) with long short-term
memory. Appl Thermal Eng 115253
27. Liu S (2019) Wind/solar power forecasting using improved LSTM neural networks. Diss
28. Hochreiter S, Schmidhuber J (1996) LSTM can solve hard long-time lag problems. Adv Neural
Inf Process Syst 9:473–479
29. Srivastava S, Lessmann S (2018) A comparative study of LSTM neural networks in forecasting
day-ahead global horizontal irradiance with satellite data. Sol Energy 162:232–247
30. Abdel-Nasser M, Mahmoud K (2019) Accurate photovoltaic power forecasting models using
deep LSTM-RNN. Neural Comput Appl 31(7):2727–2740
31. Liu H, Mi X-W, Li Y-F (2018) Wind speed forecasting method based on deep learning strategy
using empirical wavelet transform, long short-term memory neural network and Elman neural
network. Energy Convers Manage 156:498–514
32. Yu Y, Cao J, Zhu J (2019) An LSTM short-term solar irradiance forecasting under complicated
weather conditions. IEEE Access 7:145651–145666
33. Schäfer AM, Zimmermann HG (2006) Recurrent neural networks are universal approxima-
tors. In: International conference on artificial neural networks. Springer, Berlin, Heidelberg,
September 2006, pp 632–640
34. Hochreiter S (1997) JA1 4 rgen Schmidhuber. Long Short-Term Memory. Neural Computat
9(8)
35. Gaizen S, Fadi O, Abbou A (2020) Solar power time series prediction using wavelet analysis.
Int J Renew Energy Res (IJRER) 10(4):1764–1773
A Comparison Study of Machine
Learning Methods for Energy
Consumption Forecasting in Industry
Abstract The rapid expansion in the industrial sector requires good control of
energy consumption, whereas the prediction of the consumption is the most impor-
tant factor for energy management as well as environmental protection. This paper
presents a model for predicting energy consumption using ML machine learning
models such as Linear Regression (LR), Support Vector Machine (SVM), Deci-
sion Tree (DT), Artificial Neural Networks (ANN), and Recurrent Neural Networks
(RNN) are used to predict energy consumption based various attributes as inputs.
In order to evaluate the effectiveness of the proposed approaches, a simulation is
carried out using the Python software. Finally, according to simulation results, the
Linear Regression (LR) method is efficient and give better performances.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 165
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_16
166 M. Bahij et al.
2 Related Work
Generally, the ML algorithm became more used for training prediction models from
existing data compared with conventional methods. This section discusses the papers
linked to this study.
The authors in [9] develop a model for energy consumption forecasting according
to three statistical learning methods included regression trees, RL and multivariate
adaptive regression (MARS).
In paper [10], the authors propose an algorithm based on the ANN for prediction
of consumption load in the building from a set of hourly energy data measures of
consumption of 93 households for 2 months. For the forecasting of energy consump-
tion in the building or industry, multiple linear regression methods, the ANN, and
the genetic algorithms are used to process for data, in order to generate the desirable
A Comparison Study of Machine Learning Methods … 167
prediction model [11]. In article [12], the ANN used to predict the consumption in
the building for the cooling, the heat and the consumption and emission of carbon
dioxide.
In [13], the SVM method is compared with others for a tunnel settlement prediction
problem. This approach is most used in prediction problems [14]. The authors in [6]
used the improved TD as a predictive tool, where she found that this technique had
good results for a small to medium-sized dataset.
The DT presents an interpretable architecture, efficient for solving the problems
the regression and the classification, which the DT does not require a large number
of parameters and the results obtained are easy to in the classification model.
From this brief rated work, a model for energy consumption forecasting in indus-
trial depends on historical data, which comparison between the different algorithms
in order to evaluate the most efficacious methods in this study.
3 Background
The RNN is a new technique recently used in the processing of time series of energy
consumption, where their structure is developed by the ANN network as it combines
contextual information from previous entries.
According to the memory mechanism the information is transmitted to the network
one by one and the network nodes store their state at a time stage and use it to inform
the next stage [5]. The {x1, x2,…, xn} input sequence is modeled at the RNN level
using Eq. 1 looping:
h i = f (h i−1 , xi ) (1)
According to the three layers are RNN, the calculation of each layer is described
in detail below:
a Input layer
Input layer: The word sequence of the text at the input layer is of the sequence
{x1 , x2 . . . . . . .xn },where the current input vector mapping is {x1∧ , x2∧ . . . . . . .xn∧ }.
b Hidden layer
∧
At the input layer the vector of the sequence xn−1 , as well as h t−1 describe the hidden
state at the previous moment of the output vector. The output of the cache state is
expressed according to Eq. 2:
∧
h = σ W hx xn−1 + W hh hn−1 + bh (2)
168 M. Bahij et al.
Where σ means the activation function of each element, and bh presents the vector
bias of the hidden layer.
c Output layer
The output layer compute is based on the current cache layer memory state expression
as input. Equation 3 describes the calculation process:
yt = so f tmax U y h t + by ∈ R V
P(xn /x1 , x2 . . . . . . .xn−1 ) (3)
From where yt presents the probable output displayed, and U y h t defines the weight
matrix between the hidden state and the output layer.
In the ML field, neural networks are widely known as an energy estimation tool in
building and industry. According to their strong learning ability the ANN are more
used to solve the non-linear problem [4].
The ANN composed of the following layers; the input layer contains 4 neurons
reserved for attributes, and the hidden layer is composed of 10 neurons, finally, 1
neuron in the output that expresses the prediction of energy. Figure 1 shows the
Model ANN. The construction of the ANN algorithm according to six steps;
• Preparation of the systems data base.
• Initialization of the weights.
• From the input value of previous neurons calculates the desired output value.
• Improve the output value by acting on the average square root factor between the
output components.
• Data training phase using back-propagation algorithm.
• Forecasting model formulation.
be
Data input
Energy Consumption
forecasting
bs
Output layer
where ∗, ∗, w, ϕ(xi), and b ∈ R, describe the dot product, the weight, the nonlinear
mapping function, and the bias.
Y = ϕ0 + ϕ1 b1 + ϕ2 b2 + ϕ3 b3 . . . . . . . + ϕ P b p + θ (5)
where the function Y presents the output variable, ϕi for the regression parameters,
bi presents inputs values, and θ is the random error.
Decision trees are widely proposed as the optimal solution for classification prob-
lems, or the architecture of the decision tree provide good flexibility, efficiency and
precision for the generation of the prediction model [6].
In fact, their mechanism allows data to be partitioned using a flow chart to predict
desirable consumption. For a large amount of data, the DT is suitable to discover
hidden rules, but in classifiers it may be unstable in relation to the noise present in
the learning data.
When learning the model, the data processing undergone a process has a root node
and 2 branch nodes, in order divided into the other sub-nodes so that the final result
170 M. Bahij et al.
of the DT approach is shows the terminal node. In the paper, this method is used to
predict the outcome from the database provided by following certain decision rules.
In this work, the database established during training of the prediction model consists
of the assigned disputes including temperature, humidity, lighting and consumption
at the production level.
The overall diagram proposed system is described in Fig. 2. The data set estab-
lished in the model formation consists of two parts, one for training and the rest for
testing, this database undergoes such a process of normalization and standardization
described in Eq. 6:
x−μ
x= (6)
σ
Where μ describes the mean of characteristic vectors, σ this standard deviation,
x is the original characteristic vector and x is the vector of characteristics after
normalization. Based on the difference between the expected results and the actual
data, the efficacy of the approach used is determined.
ML Approche Model
Estimation
In this part, the simulation and analysis results of RNN, ANN, SVM, DT and RL algo-
rithms are expressed. The approaches are designed in the Python environment with
the PyTorch tensor library. Then, to examine actual and predicted energy consump-
tion using ANN, RNN, DT, SVM, and LR approaches are shown in Figs. 3, 4, 5, 6
and 7 figures respectively.
The performance of these algorithms used in the theme of precision are of average
absolute percentage Error (MAPE) described in Eq. 6:
n
100% y − yi∧
M AP E = (6)
n I =0 yi
The energy consumption profile is defined by the correlation between the actual
and predicted energy consumption presented in Fig. 3. At the level of the simulation
using the ANN approach, we can see that the forecast curve accurately follows the
energy consumption compared to the real consumption. Then, the forecasting error
is less than 0.2.
In the RNN, the prediction error is of order 0.9, which means that this algo-
rithm is less efficient than the ANN. Figure 4 expresses the behaviour of the energy
consumption between the measured real and the profile estimated by the model.
The performance of the SVM is demonstrated in Fig. 5, where the error value
does not exceed 0.05 of the references, that expresses that the SVM has a strong
training capacity. In Fig. 5, it is well defined that the estimated energy consumption
is close to the curve of the real consumption.
For DT, the error is 0.5, which the variation in the energy consumption profile
between the actual consumption and the estimated is shown in Fig. 6
Figure 7 expresses the forecast of the energy consumption profile which is similar
to the actual consumption. The error obtained is of reference value 0.6 × 10−3 in the
LR approach.
The performance criteria used in the comparison of the five approaches are
the following: the absolute percentage of error MAPE, the training score and the
prediction score are describing in Table 1.
Moreover, The LR training score is 100% implies that higher performance algo-
rithm. The score is 99.999% in the SVM and the same in DT method. But those are
less inferior to the first RL algorithm.
The RNN has a value of 99.708% and the ANN has a score of 99.506%, which
presents the least efficient algorithm in the prediction of the model.
In Similarly, the value of the mean square error is 4.202 for ANN, which is lower
than the RNN approach of the order of 1.56. The error is 5.75 10–2 in DT method.
Then, the SVM is the value of 5.05 10−4 which expresses that the ANN and RNN
algorithms perform less than the two methods preceding. The LR most effective
approach in model training with a lower order error 3.45 10−28 than all approaches
proposed.
5 Conclusion
References
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 177
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_17
178 M. Timmi et al.
1 Preliminaries
This article discusses a study related to the field of data mining and how it can be
diligently implemented within the educational field. Besides, it expands upon the
central features of educational data mining processes and the objectives of these
operations, along with some rudiments that pave the way for an effective application
of data mining algorithms.
Historically, data mining is not novel, because it mainly relies on statistics that
date back to the middle of the 18th century. The current state of data mining is the
result of substantial efforts that have been accumulated for many years [1], yet it is
worth noting that this area has witnessed a remarkable development recently.
Returning to the roots of educational data mining, we find that research in
this area is relatively new. Although research and analysis of data, resulting from
educational software, started a long time ago, knowledge exploration in educa-
tional data mining (EDM) has been referred to, in various conferences and scien-
tific journals, as a separate field itself in recent times. This claim has also been
advocated in a wide range of reputable conferences on educational technologies,
namely the international conferences revolving around the following matters: educa-
tional data mining/artificial intelligence in education/intelligent systems/user
modeling, adaptation and personalization (UMAP).
Educational data mining is defined in a variety of ways. However, the one that is put
forward in the Journal of Educational Data Mining is fundamentally insightful. In
this sense, the journal regards EDM as the process whereby methods are developed
to explore idiosyncratic data that come from educational environments [2], and the
use of these methods to deepen our own understanding of learners, and to reflect
them on the environment in which they operate. Correspondingly, another definition
considers EDM as the use of information mining techniques for specific types of
data extracted from educational environments to either deal with or solve important
education related issues. As far as we can tell, both definitions focus on extracting
knowledge from educational data to improve education systems [3]. The exploration
of knowledge in the field of education can occur by means of piecing the following
elements together: computer science, education and statistics [4]. With that in mind,
the intersection of the previous three areas forms other sub-areas, such as learning
analytics, e-learning and data mining [5] (Fig. 1).
Educational data mining is a multidisciplinary arena, for it uses various methods
and techniques, namely statistics, data mining, information retrieval systems, sugges-
tion systems, machine learning, science education, teaching principles and others.
Notwithstanding, the process of selecting these techniques depends predominantly
on the studied educational problem [6].
The Review of Objectives, Methods, Tools, and Algorithms … 179
The stages through which the process of exploration in educational systems goes
can be diagrammed as follows[7]: (Fig. 2).
It is also worth mentioning that there are other areas where data mining is applied,
such as criminal investigations, industrial engineering, fraud detection and customer
relationship management.
3 Selection of Features
The selection of features consists in choosing a small subset of features that are
prerequisite to describe the target concept by means of prediction, classification,
regression or clustering …. (It depends on the method used to explore the educa-
tional data). The selected subset is chosen with the assumption that the input data
contains a range of irrelevant features, and its removal would not affect in any way
the accuracy of the classification model (of the method used), as a matter of example.
Obviously, this is advantageous in the sense that it allowed a remarkable reduction in
computational complexity, better interpretability of the model and lessened data fit
180 M. Timmi et al.
4.1 Learners
Providing notes and advice to students, meeting the needs of the students, working
to increase the students’ productivity and academic level, etc.
4.2 Teachers
Expand their understanding of the students’ nature, the progress of the educational
process, and get them to update their teaching methods and performance.
4.3 Researchers
Develop and compare data extraction techniques to intensify their focus on optimal
suggestions for each problem or specific pedagogical tasks, assess the effectiveness
of learning when using different methods, environments. etc.
4.4 Administrators
Come up with alternative canals to arrange resources (human and material) and assess
their educational establishments.
The Review of Objectives, Methods, Tools, and Algorithms … 181
It is also important to admit that the process of involving all stakeholders and
providing them with the results of the analysis is a critical task that should be consci-
entiously managed. Actually, some of these interests may end up taking an exclu-
sively different turn. Hypothetically, the use of learning analytics (LA) by personnel
managers to introduce technology into teaching methods might be misperceived, as
teachers would regard that as an attempt to evaluate or judge their teaching methods.
Additionally, the same goes for students who may feel uncertain to use personal data
for their assessment and classification. As far as we can see, this can produce an
unexpected consequence; teachers and students alike will be unwilling to use these
new technologies, or to partake in the various stages of the analysis.
Previously, the objectives were classified based on the types of end users. However,
it is challenging to categorize the objectives of knowledge exploration integrally
in education systems using one criterion, particularly, when all the objectives are
linked to several users. In this regard, another trustworthy research has described the
classification of exploration objectives in educational data as follows [9]:
• Student modeling: user modeling in the education system requires providing a
detailed account about the characteristics of students as well as their situation.
It is primordial to have an all-inclusive understanding of their knowledge, skills,
experiences and motivations. On the other hand, one needs to examine certain
types of problems, which can negatively derail the educational process, namely
the use of ineffective educational resources. Consequently, the objective is to
create or improve the student’s model based on the available information.
• Observe the students’ level and their results: the aim is to anticipate the outcomes
of every educational procedure by measuring how active the interactions in the
classroom are.
• Search for new learning suggestions: the aim is to suggest new content and better
tasks that are in conformity with the students’ current level/situation.
• A profound exploration of the learner’s behavior: This objective can actually vary.
Based on the previously delivered answers, the classification of students has to
make recourse to numerous factors, namely their personal profiles.
• Communicate with stakeholders: the objective is to assist teachers and get
managers to analyze the activities accredited to students.
• The analysis of the terrain structure: the goal is to define the terrain structure to
know the optimal content that can be added to the material and choose the best
possible sequence for the educational process. This can be achieved by devel-
oping the skill of predicting the student’s performance and determine the ideal
educational environment that would guarantee a fulfilled learning experience.
• The Improvement of educational content: it is closely linked to the two previous
objectives. The aim here is to decide on how to improve the educational material
(content, activities, etc.) using student information.
• Study the effects of educational support that is put forward by education systems.
It is noteworthy that the previously named objectives of educational data mining
are meant to improve and computerize education systems. More significantly, it can
182 M. Timmi et al.
be proclaimed that student modeling is the main key to achieve many objectives and
tasks through the exploration of educational data [7].
5.1 Prediction
The goal of this method is to develop a model that allows the inference of a particular
aspect of this data (called a predicate variable) in a set of aspects within the same data
reservoir (called predictive variables). The applications of educational data mining
using this method attempt to anticipate the student’s academic success by reviewing
their behavior in the classification-based virtual environment. Besides, going about
predicting the results of students who have completed the same teaching unite is
manifestly useful to back up those who may fail to fulfill it. The KNN algorithm
(K nearest neighbors) is among the techniques that are highly recommended in this
respect. It should be pointed out that the algorithms used in this area, are virtually
expressed in the interactions of students with the available sources in the form of
multidimensional rays [12, 13].
5.2 Clustering
This method implies, in the fewest possible terms, that the classified samples which
encompass data are naturally similar to each other. A cluster is widely defined as a
group of identical objects that differ from those found in other clusters. It is broadly
agreed upon that the efficiency of the cluster upsurges suggestively depending on
how similar or different it is from the elements contained in other clusters. This
mechanism is postulated to discover similar features between different behavioral
patterns manifested by students [13].
It refers to the examination of variables that constitute a data sample. This connection
can come in different forms, such as association rules mining, sequential pattern
The Review of Objectives, Methods, Tools, and Algorithms … 183
mining, correlation mining, and casual data mining. In the realm of educational data
mining, “relationship mining” is manipulated to determine the possible connections
between students’ online activities and their grades [13].
5.4 Distillation
This method includes portraying data in a manner that allows the user to rapidly
detect and categorize the data features. This technology chiefly depends on data
Summarization, visualization, and the use of Interactive Interfaces to demonstrate
useful information as well as supporting decision-making [13].
This method ruminates on the discovery of certain data segments with a marked
difference, compared to other casual genres of samples. It is predominantly applied
to explore educational data in order to reveal common behaviors displayed by the
involved stakeholders. Besides, it has been proven a practical medium to spot students
with learning anomalies [13].
The structural analysis of this method is looking forward to studying the relationship
between individuals rather than considering their behavioral characteristics. From
an empirical viewpoint, social relationships are merely transcribed contractually to
establish multiple social links. This elementary segment can manifest itself, in terms
of educational data exploration, to interpret and analyze relationships in collaborative
assignments [9, 12, 13].
This method sheds light on extrapolating useful styles, patterns, and rules from
unstructured texts, namely HTML files, discussion messages, and emails. The Tasks
comprise text categorization, text compilation, analyzing the emotional insights
within a text, synthesizing documents and modeling relationships between diverse
objects. Essentially, the main reason behind using text mining is to examine the
content of discussions, forums, chats, web pages, documents… etc. [12, 13].
6.1 K-means
“K-means” is a reiterative algorithm that attempts to divide the dataset into K pre-
determined and non-overlapping subgroups (clusters). Here, every data point is only
affiliated to one group [15]. The pseudocode of K-means algorithm as follows [16]:
The Review of Objectives, Methods, Tools, and Algorithms … 185
6.2 Apriori
Apriori algorithm is given by R. Agrawal and R. Srikant in 1994 for finding frequent
set of items in a dataset for Boolean association rule. Surprisingly enough, this
significant appellation was given to the algorithm for its reliance on prior knowledge
of frequent itemset properties [15, 17].
6.3 EM Algorithm
Google Search famously adopts PageRank algorithm to categorize web pages in their
search engine results [19].
Rapid Miner is widely known for being a readymade, open source, no-coding required
software, which helps in giving advanced analytics [11].
7.2 Weka
Weka is a range of machine learning algorithms for data mining tasks [11, 15].
7.3 Orange
Orange-This tool comes with visually programmed environments. It deals with tools
of data importing, data dragging and dropping widgets, as well as providing links to
connect dissimilar widgets together so that the workflow can be completed.
7.4 R
7.5 Knime-Primarily
Knime-Primarily used for data preprocessing. It is a powerful tool with GUI that
shows the network of data nodes.
The Review of Objectives, Methods, Tools, and Algorithms … 187
7.6 Tanagra
Tanagra is a free open source of data mining software put in place for mere academic
and research purposes.
7.7 XL Miner
XL Miner-It is the only comprehensive data mining add-in for Excel, with neural
nets, classification and regression trees, logistic regression, linear regression, Bayes
classifier, K-nearest neighbors, discriminant analysis, association rules, clustering,
principal components, and a lot more.
8 Conclusion
Educational data mining (EDM) has shown itself as an enormously fertile research
area, considering its remarkable potentials to contribute to a deeper understanding of
teaching, learning and motivational processes in individual and collaborative teaching
environments.
Currently, EDM’s main contributions are focused on two main axes:
• Data analysis and creation of models to sharpen the collective understanding of
the learning process.
• Developing more effective and innovative methods to support learning and
enhance the qualitative dimension of using educational web applications (E-
Learning, M-Learning).
Today, around the world, many intelligent tutoring systems use EDM techniques
to provide more personalized and better learning.
References
1. Şahin M, Yurdugül H (2020) Educational data mining and learning analytics: past. Present
Future 9:121–131
188 M. Timmi et al.
2. Baker RS, Yacef K (2009) The state of educational data mining in 2009: a review and future
visions. JEDM| J Educ Data Min 1(1), 3–17
3. Romero C, Ventura S (2017) Educational data science in massive open online courses. Wiley
Interdisciplinary Rev Data Min Knowl Discov 7(1):e1187
4. Romero C, Ventura S (2020) Educational data mining and learning analytics: an updated survey.
WIREs Data Min Knowl Disc. https://doi.org/10.1002/widm.1355
5. Salloum SA, Alshurideh M, Elnagar A, Shaalan K (2020) Mining in educational data: review
and future directions. In: Joint European-US workshop on applications of invariance in
computer vision. Springer, pp 92–102
6. Romero C, Ventura S (2020) Educational data mining and learning analytics: an updated survey.
WIREs Data Min Knowl Disc 10(3):e1355. https://doi.org/10.1002/widm.1355
7. Anjum N, Badugu S (2020) A study of different techniques in educational data mining. In:
Advances in decision sciences, image processing, security and computer vision. Springer,
Cham, pp 562–571
8. Acharya A, Sinha D (2014) Application of feature selection methods in educational data mining.
Int J Comput Appl 103(2)
9. Mitrofanova YS, Sherstobitova AA, Filippova OA (2019) Modeling smart learning processes
based on educational data mining tools. In: Smart education and e-learning. Springer, pp 561–
571
10. Peña-Ayala A (2014) Educational data mining: applications and trends
11. Kumar AD, Selvam RP, Kumar KS (2018) Review on prediction algorithms in educational data
mining. Int J Pure Appl Math 118(8):531–537
12. Bakhshinategh B, Zaïane O, Elatia S, Ipperciel D (2017) Educational data mining applications
and tasks: a survey of the last 10 years. Educ Inf Technol 23. https://doi.org/10.1007/s10639-
017-9616-z
13. Aldowah H, Al-Samarraie H, Fauzy WM (2019) Educational data mining and learning analytics
for 21st century higher education: a review and synthesis. Telematics Inform 37:13–49. https://
doi.org/10.1016/j.tele.2019.01.007
14. Shende A, Thakare MS, Byagar S, Joshi MAN (2020) The review of different data mining tools,
techniques and algorithms for the data mining in education. Our Heritage 68(27):290–296
15. Dhingra K, Sardana KD (2017) Educational data mining: a review to its future vision. Int J
Technol Transfer Commer 15(3):309–321
16. Olukanmi P, Nelwamondo F, Marwala T (2019) Rethinking k-means clustering in the age of
massive datasets: a constant-time approach. Neural Comput Appl 1–23
17. Bhandari A, Gupta A, Das D (2015) Improvised apriori algorithm using frequent pattern tree
for real time applications in data mining. Procedia Comput Sci 46:644–651. https://doi.org/10.
1016/j.procs.2015.02.115
18. Kurban H, Jenne M, Dalkilic MM (2017) Using data to build a better EM: EM* for big data.
Int J Data Sci Anal 4(2):83–97
19. Agryzkov T, Curado M, Pedroche F, Tortosa L, Vicent JF (2019) Extending the adapted
pagerank algorithm centrality to multiplex networks with data using the Pagerank two-layer
approach. Symmetry 11(2):284
20. Wati, M, Budiman E, Kridalaksana AH, Haviluddin H, Dengen N, Purnawansyah P (2018)
Performance of decision tree C4. 5 algorithm in student academic evaluation
Brain-Computer Interface: A Novel EEG
Classification for Baseline Eye States
Using LGBM Algorithm
Abstract This work has been carried out in the development of the best BCI systems
to recognize EEG visual system signal helps tired drivers, the elderly, patients, and
the disabled to drive on good safety condition. The method used in this article based
on an application of rhythms (delta, theta and alpha) on EEG acquisition channels to
increase the accuracy of the prediction system, after using delta rhythms on selected
electrodes using ExtraTrees algorithm, we notice an increase in accuracy from 95.46
to 99.97%, also this system has a good prediction speed of 12560 samples per second
during testing. Therefore, instead of minimizing the number of acquisition electrodes,
these bad electrodes can be used to increase the performance of the BCI system using
rhythms (delta, theta and alpha).
1 Introduction
The brain computer interface systems were launched in 1924 with the first acqui-
sition of EEG signals by Hans Berger. Each organ connects to a specific area in
the brain via the neural system, the distribution of electrical activity in the brain
refers to a sensation or movement of an organ, so these signals transmit through a
neurons network between organ and its analysis area in the brain [10]. The aim is to
provide a direct link between observed human intentions in the brain activities and
computer control, an important feature of current BCI systems is the high complexity
of their functional extracts compared to their simple classifications (usually linear)
[8]. In the human brain, there are billions of neurons and these neurons commu-
nicate with each other using weak electrical signals, using these electrical signals
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 189
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_18
190 S. Abenna et al.
brain activity is measured and these signals are called EEG signals [2], the main
purpose of using EEG signals is to diagnose brain diseases [3], most medical appli-
cations are based on BCI systems, because it can be used to prevent harmful effects
of smoking, alcohol consumption, it also helps detect and diagnose the brain disor-
ders, sleep disorders and tumours, the BCI can also be used to restore the health of
people with a cerebral vascular accident, disability and mental disorders [7]. The
EEG signals is recorded by several electrodes placed on specific areas from scalp,
the EEG signals have a distinctive high resolution of up to one millisecond, which is
still not possible with the latest imaging technologies such as computerized tomog-
raphy (CT) or magnetic resonance imaging (MRI) [13]. The EEG technology has
been progressively used in the auxiliary diagnosis of diseases such as schizophrenia,
mild cognitive impairment, epilepsy and Alzheimer’s disease. Above all, a close
relationship has been found between brain and depression, a cognitive ability of
depressed patients changes with mood, and these changes affect EEG [4], the diffi-
culties required in EEG signals cannot be controlled because many activities take
place in the brain at same time, the use of adaptive factors ap-pears to be beneficial
for biological symptoms such as electroencephalogram [11], there have been many
EEG classification studies in recent years, these studies used different classification
techniques, collected their performance and evaluated different collections of char-
acteristics, among these categories, k-nearest neighbor (k-NN), linear discriminant
algorithm (LDA), support vector machine (SVM), artificial neural network (ANN)
have become popular [6, 11]. In this paper, we have been interested in developing a
prediction system using EEG signals for eye states (open or closed), such that this
system is based on three main steps as a sequel, the selection of electrodes that allows
to determine the most important electrodes for classification and extraction of char-
acteristics based in the application of delta rhythms, theta and alpha on selected bad
electrodes, the final step is the use of the LGBM algorithm to classify EEG signals.
As a maximum accuracy value obtained in this work is in the order of 99.97%. The
rest of this article is organized as follows, Sect. 2 present the data used, feature
extraction and classifier proposed for system, the results obtained and discussed are
presented in Sect. 3, while Sect. 4 provides conclusions and an overview of future
work.
2 Methods
2.1 Dataset
This paper uses statistical data from an acquisition of EEG signals for eye states
(open or closed) using 14 non-invasive electrodes (Fig. 1a) from UCI Database [15].
The signals acquired using Emotive EEG Neuroheadset device use the labels ‘0’ to
indicates when eyes are open and ‘1’ to indicates when eyes are closed, the total
Brain-Computer Interface: A Novel EEG Classification … 191
Fig. 2 Proposed general architecture for baseline eye states recognition system
number of data is 14980 samples, such as training data use 80% of the total data and
the rest 20% is used as testing data (Fig. 2).
This step is very important in determining the electrodes that are weak system recog-
nition accuracy. There are many feature selection algorithms, in this work we used
ExtraTrees (ET) [12] based on SelectKBest and chi2 algorithms to select important
channels.
2.4 Classification
The classification of EEG signals is a very important step to resume the quality of
the BCI system, from many algorithms, the LGBM algorithm was chosen because of
its effectiveness. Like LGBM (Light-Gray Box Model) is a gradient-enhancing tool
that uses leaf tree growth for learning, which speeds up the formation and accuracy
of better results, even when you use less memory [9].
The principle measures are used to evaluate the performance of prediction in this
paper are:
With TP, FP, TN and FN are calculated from the confusion matrix of prediction
during testing (Fig. 3b):
-TP (True Positive): refers to a case where an alarm is raised when eyes open.
-TN (True Negative): refers to a case where an alarm is not raised when eyes
closed.
-FN (False Negative): refers to a case where an alarm is not raised when eyes
open.
-FP (False Positive): refers to a case where an alarm is raised when eyes closed.
Fig. 3 Evaluation metrics (a) and confusion matrix for binary class (b)
Brain-Computer Interface: A Novel EEG Classification … 193
3 Results
The results of this paper were obtained using python3 on a core i-5 computer with
4-CPU and 6-Gb in RAM. Figure 4a illustrates a classification of electrodes from the
bad electrode used (P8) to the best electrode (O1), the Fig. 4b shows the correlation
values between the electrodes. we notice the existence of links between different
electrodes of each group (AF3, P8, F8), (F7, F3), (FC5, O1), (P7, AF4) and (O2, T8),
compared to Fig. 1b, each set of electrodes belongs to the same nerve which implies
the same signal characteristics and ‘1’ as correlation value between the electrodes,
which shows the good quality of the signals acquired by Emotive EEG Neuroheadset
device.
The classification of EEG data is the most important step in evaluating results. For
this, a set of metric values such as accuracy (ACC), sensitivity (SEN), specificity
(SP), precision (PR), kappa-score (k), F1-score (F1), mathews correlation coeffi-
cient (MCC), mutual information (MI), zero on loss (ZOL), pretreatment time (Ta),
classification time (Tc) and prediction time (Tp) were used.
Table 1 illustrates the classification results using the LGBM algorithm and rhythms
on the selection channels, this table shows the increase in classification performance
when applying the rhythms on several electrodes, so the classification time remains
constant, in such a way the accuracy values increased from 95.46% when not using
Table 1 Classification results with LGBM
194
feature extraction to 99.97% when using feature extraction. Compared to the results
with literature in Table 2, we notice that the wrong method was used by [1] using
the Delta-Theta-SVM method with an accuracy value of 70.6%, on the other hand at
our work, the accuracy value takes 99.97% and the specificity takes the ideal value
(100%). So to calculate the speed of prediction of the system during testing, one
divides the total number of test data over the duration of pretreatment and prediction,
to calculate the total number of test data from Table 1, we divide the value of the
ZOL on the error rate (100-ACC (%)), also the preparation time and prediction can
be calculated by a simple addition between 20% of Ta and Tp (0.2 * Ta + Tp). After
calculating the average speed of system prediction, it is found that its value takes
12560 samples per second.
Every literature work suffering some technical deficiency to not find good results,
the Zhou et al. [15] method use Delta rhythms only in the stage of feature extraction
of EEG data as in our work, but these rhythms can delete an important set of data
for classification, so we must add the rhythms Theta bandwidth as in the method of
Seghafi et al. [1] so the filter passes band uses bandwidth between 0.5 and 7 Hz to
add more character to the training data, with the possible minimum of noise, we also
notice the possibility of improving the quality of classification when merging two
algorithms as in the case of the Hamilton et al. [5] method, such that we notice that
Brain-Computer Interface: A Novel EEG Classification … 197
the accuracy value increases by 95.10% and 97.30 to 97.40% for this UCI dataset
classification, so we notice that RRF and LGBM finds the same accuracy value
without feature extraction algorithm, which shows that this work makes a good
correction of the different methods of literature.
4 Conclusion
In conclusion, this work improves the recognition of eye states, such as accuracy takes
the value 99.97% using the LGBM algorithm and low rhythms of EEG signals from
bad electrodes selected for the feature extraction, and compared to other methods,
we notice that our system is reliable with a high average prediction speed of 12560
samples per second. We hope that this study helps other researchers develop good
BCI systems to help and improve human life. For the future work, we will now
improve a new method for the feature extraction and classification steps of motor
movement-imagery recognition.
References
1. Saghafi A, Tsokos CP, Goudarzi M, Farhidzadeh H (2017) Random eye state change detection
in real-time using EEG signals. Expert Syst Appl. https://doi.org/10.1016/j.eswa.2016.12.010
2. Khosla A, Khandnor P, Chand T (2020) A comparative analysis of signal processing and
classification methods for different applications based on EEG signals. Biocybernet Biomed
Eng 5–6. https://doi.org/10.1016/j.bbe.2020.02.002
3. Emrah A, Turker T, Sengul D (2020) A Tunable-Q wavelet transform and quadruple symmetric
pattern based EEG signal classification method. Medical Hypotheses 1–2. https://doi.org/10.
1016/j.mehy.2019.109519
4. Cai H, Qu Z, Li Z, Zhang Y, Hu X, Hu B (2020) Featurelevel fusion approaches based on
multimodal EEG data for depression recognition. Inf Fusion 1–2. https://doi.org/10.1016/j.inf
fus.2020.01.00
5. Hamilton CR, Shahryari S, Rasheed KM (2015) Eye state prediction from EEG data using
boosted rotational forests. In: 2015 IEEE 14th international conference on machine learning
and applications. https://doi.org/10.1109/ICMLA.2015.89
6. Nicolae C, Nirvana P, Marius L (2020) Reading into the mind’s eye: Boosting automatic
visual recognition with EEG signals. Neurocomputing. https://doi.org/10.1016/j.neucom.2019.
12.076
7. Edla DR, Ansari MdF, Chaudhary N, Dodia S (2018) Classification of facial expressions from
EEG signals using Wavelet Packet Transform and SVM for wheelchair control operations.
Procedia Comput Sci 1–2. https://doi.org/10.1016/j.procs.2018.05.081
8. Farquhar J (2009) A linear feature space for simultaneous learning of spatio-spectral filters in
BCI. Neural Networks J. https://doi.org/10.1016/j.neunet.2009.06.035
9. Ke G, Meng Q, Finley T, Wang T, Chen W, Ma W, Ye Q, Liu TY (2017) LightGBM: a highly
efficient gradient boosting decision tree. In: 31st conference on neural information processing
systems, Long Beach, CA, USA
10. Eitan N, Alex F, Dan F (2020) Real-time EEG classification via coresets for BCI applications.
Eng Appl Artif Intell. https://doi.org/10.1016/j.engappai.2019.103455
198 S. Abenna et al.
11. Parvinnia E, Sabeti M, Zolghadri Jahromi M, Boostani R (2014) Classification of EEG Signals
using adaptive weighted distance nearest neighbor algorithm. Comput Inf Sci 1–2. https://doi.
org/10.1016/j.jksuci.2013.01.001
12. scikit-learn user guide, Release 0.20.4, 2019, pp 255–257
13. Amin SU, Alsulaiman M, Muhammad G, Mekhtiche MA, Shamim Hossain M (2019) Deep
Learning for EEG motor imagery classification based on multi-layer CNNs feature fusion.
Future Gen Comput Syst 1–2. https://doi.org/10.1016/j.future.2019.06.027
14. Reddy TK, Behera L (2016) Online eye state recognition from EEG data using deep architec-
tures. In: 2016 IEEE international conference on systems, man, and cybernetics. https://doi.
org/10.1109/SMC.2016.7844325
15. Zhou Z, Li P, Liu J, Dong W (2019) A novel RealTime EEG based eye state recognition system.
Springer. https://doi.org/10.1007/978-3-030-06161-6 17
A Light Arabic POS Tagger Using
a Hybrid Approach
1 Introduction
The world of the Internet has known a huge and continuous growth regarding the
Arabic content in terms of texts, videos, images, etc. Statistics show1 that this content
has increased by 9348% between the year of 2000 and 2020. As a result, developing
Arabic Natural Language Processing (ANLP) tools is required to process this huge
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 199
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_19
200 K. Tnaji et al.
and growing content. In fact, there are many applications that have been already
developed for Arabic such as Part-of-speech (POS) taggers. This latter is an essential
tool in many NLP applications such as word sense disambiguation [18], question
answering [11] and information retrieval [9].
However, the number of Arabic POS taggers that can be used as rapid and free
tools is still low [1, 7, 8]. In addition, they suffer from three major problems. The
first and foremost problem concerns words having more than one possible tag when
they are used in different contexts. The second problem concerns words not found
in the annotated training set since the majority of the POS taggers can tag only
words that actually exist in the training set. Therefore, these unknown words lead
to a major problem in any tagging system and always decrease its performance.
Finally, as will be detailed in the literature review, Arabic researchers use different
and non-standard tag sets such as [4, 10, 14]. This situation leads to the difficulty of
comparing or combining these POS taggers.
Therefore, our objective is to develop a new tool for Arabic POS tagging fixing the
above-mentioned drawbacks. This POS tagger is developed using a hybrid approach
combining a statistical one to deal with ambiguous words and a symbolic one to
cope with unknown words. Finally, instead of specifying a full and very detailed tag
set, we limit our tag set to four generic tags (noun, verb particle and punctuation)
since they are enough in many kinds of applications. We consider then our new POS
tagger as a light Arabic POS tagger.
The rest of this paper is organized as follows. The next section presents related
works concerning POS tagging techniques for unknown and ambiguous words.
Section 3 provides the steps for developing our POS tagger. The results of experi-
ments are described and commented in Sect. 4. The exploitation of this system is
described in Sect. 5. Finally, some conclusions and future works are given in Sect. 6.
2 Related Works
Research on POS tagging has a long history where numerous techniques have been
successfully applied. Some of them focus on the use of linguistic rules. Others
consider POS tagging as a classification problem: the tags are the classes and a
machine learning algorithm is used to assign each occurrence of a word to one class
based on the evidence from the context.
Different Arabic taggers have recently been proposed. Some of them use super-
vised machine learning approaches such as [1] while others use hybrid approaches,
combining the rule-based approach and the statistical one such as [2]. Some of these
POS taggers handle unknown words such as [13]. A detailed review about the most
recent works related to Arabic POS taggers can be found in [16].
The major problem in all these Arabic POS taggers is the absence of a standard tag
set. For example, the number of tags varies between 6 tags in the case of Columbia
Arabic Treebank POS tag set [10], and 2000 tags in the case of Penn Arabic Treebank
(PATB) [14]. This situation makes it impossible to compare or combine these existent
A Light Arabic POS Tagger Using a Hybrid Approach 201
systems. A comparison about some of the existing Arabic tag sets can be found in
[5] where we can notice the big difference between the tag sets because each one is
developed for a specified application.
Some studies have been done to develop an Arabic standard tag set such as [5, 12]
that aims to design detailed hierarchical levels of the Arabic tag set categories and
their relationships.
Relating to the rapidity of the Arabic POS tagging systems, few works report on
this measure despite its importance. Among the systems that have a known execution
time, we find farasa POS tagger [7] which is able to process 2000 words per second.
3 Our Approach
Let us remind that our objective is to develop a POS tagger with unknown words
handling. To avoid the non-standardization of the tag sets and to improve the POS
execution time, one solution consists in reducing the tag set containing the basics
{noun, Verb, Particle, Punctuation}.
The development of this POS tagger is based on a supervised machine learning
approach exploiting an annotated corpus. This system is developed using the Hidden
Markov Model (HMM) combined with a decision tree guessing model based on word
substring information.
The data set used for this POS tagger is the NEMLAR corpus. This corpus was
produced within the NEMLAR project2 [3] and consists of about 500,000 words of
Arabic text from 13 different categories, aiming to achieve a well-balanced corpus
that offers a representation of the variety in morphology, syntactic, semantic and
pragmatic features of modern Arabic language such as political news, political debate
or Arabic literature.
The tag set existing in the NEMLAR corpus is detailed, and since we typically
need only the 4 elementary tags, we performed a simple modification on this corpus
to map the used NEMLAR tag set to our four basic tags. For instance, adjectives,
nouns and proper nouns are all mapped to the basic noun tag.
2 https://www.nemlar.org
202 K. Tnaji et al.
3.2 Processing
A POS tagging system provides the most likely tag for each word of the sentence. Our
system is based on the hidden Markov model and the Viterbi algorithm. The HMM
[15] is a sequence model. A sequence model or sequence classifier is a model whose
job is to assign a tag or class to each unit in a sentence, thus mapping a sequence
of words to a sequence of tags. An HMM is a probabilistic sequence model: given
a sequence of words, it computes a probability distribution over possible sequences
of POS tags and chooses the best tag sequence by using the Viterbi algorithm.
However, HMM can only tag the words existing in the training data. To tag the
unknown words we have to pursue another technique. Indeed, our approach is to
exploit the morphology of words (word substring information) to associate with each
one its most likely tag. For instance, in the Arabic word. “( ”متفاهمونmtfAhmwn)3
[6], the three characters (‘( ’فf), ‘( ’هh)) and (‘( ’مm)) are root radical characters.
Therefore, they can be replaced by other three Arabic characters such as {‘( ’قq),
‘( ’دd), ‘( ’صS)} to produce the other Arabic word “( ”متصادقونmtSAdqwn) which
have different meaning but both words are nouns. The other characters at the begin-
ning and the end of the word represent the augmented characters. Linguistically
speaking each augmented character can be only one of these Arabic ten characters
{‘ ( ’اA), ‘( ’هh), ‘( ’یy), ‘( ’نn), ‘( ’وw), ‘( ’مm), ‘( ’تt), ‘( ’لl), ‘( ’أO), ‘( ’سs)}.
The augmented characters, also known as affixes, associated with their position in
the word play a critical role in determining the possible POS tag of the word. For
example, Arabic words “( ”یتصالحytSAlH), “( ”یتصاعدytSAEd), “( ”یتعاملytEAml) all
have three augmented characters {‘( ’یy), ‘( ’تt)} and {‘( ’اA)} in the first, second
and the fourth position and all these words are verbs. Therefore, our objective is to
extract automatically these classification rules that allow classifying a word from its
augmented which represents the affixes4 . The extraction of these rules is done by
using a decision tree algorithm. To build our decision tree algorithm.
The idea is simple and consists in building a decision tree by learning all over the
corpus and using as an attribute vector containing the prefix (first and second letters),
the suffix (last and before last letter) and the size of the word which replaces the infix
to avoid the complexity of extracting it. Consequently, the non-terminal nodes in our
decision tree are the attributes (prefix, suffixes or word’s size). The edges represent
the possible values for each attribute: Arabic letters or a number which represent the
size of the word. Finally, the terminal nodes are the POS tags.
A Light Arabic POS Tagger Using a Hybrid Approach 203
4 Experiments
The test is subdivided into four evaluation series where the first three use the Nemlar
corpus. In the first one, we test the efficiency of our system on the ambiguous words
using only HMM. In the second, we test our decision tree guessing model and for
the third series of evaluation, we test the full POS tagging system. Finally, we test
the POS tagging system using another corpus.
4.1 Evaluation 1
To evaluate the efficiency of our system on ambiguous words, we test the POS tagger
which depends only on the hidden Markov model. Therefore, it is necessary to use
a test data with no unknown words. Statistics and results are presented in Table 1.
Concerning words, the precision of the system is close to perfection. A precision
of 99% is very high, but it is justifiable. First, all the words exist in the vocabulary and
we are using a reduced tag set. In addition to that we can neglect the punctuation that
can never be misclassified. Therefore, the remaining 1% error is practically between
the verbs, nouns and the particles.
4.2 Evaluation 2
Despite that the POS tag of some words depends on its context in the sentence
and/or on the diacritics, this model provides an excellent accuracy.
4.3 Evaluation 3
To evaluate the full POS tagging system, we process with the train-test split technique.
In the first step we divide the full dataset into sentences. In the second step, we take
80% of the sentences for training and we keep 20% for the test. Table 3 summarizes
the statistics and results.
As we notice the number of misclassified words is extremely insignificant. This
is due to the very good quality of our model but also to the elementary tag set used.
Concerning unknown words, the number of misclassified words represents 31% of
the total unknown words that seems relatively high but technically it’s caused by the
hidden Markov model and/or the decision tree guessing model.
Finally, relating to the rapidity of this POS Tagger note that our model is extremely
rapid and could process about 4,473 words per second (with i3 processor and
4 GB/RAM) in comparison with farasa POS tagger [7] which can process about
2,000 word/second (with i7 processor and 16 GB/RAM).
4.4 Evaluation 4
In this series of evaluation, we tested our system on the WikiNews5 corpus. It contains
manually-segmented and POS tagged words from articles on seven domains: politics,
economics, health, science and technology, sports, arts, and culture. We started with
a preprocessing step that transforms the 16 used tags into our reduced tag set. Table 4
summarizes the statistics and results. For instance, the WikiNews tags noun, adjective
are mapped into the noun tag.
Once again, we notice the good accuracy and rapidity of this POS tagging system,
despite the considerable number of unknown words.
5 Exploitation
This system has been developed in java and has been integrated into the Safar frame-
work6 . This framework integrates several tools to process the Arabic language at
the morphological, syntactic and semantic levels containing parses, lemmatizers or
segmenters [17]. The framework is available both as a web interface7 and as an API.
Figure 2 shows the use of the POS tagger web interface. After selecting “Safar light
POS Tagger” from the input “Select Tagger”, the user can either tag a text by entering
it to the target zone or by directly importing a text file from the machine. The tagging
6 http://arabic.emi.ac.ma/safar/
7 http://arabic.emi.ac.ma:8080/SafarWeb/
A Light Arabic POS Tagger Using a Hybrid Approach 207
results can be visualized directly or they can be generated as an Excel, CSV or XML
file.
An example of using the POS tagger via SAFAR API is illustrated in Fig. 3. To do
so, we import the package “safar-modern_standard_arabic-basic-syntax-posTagger”.
The following code tags a text with only two instructions and then outputs the POS
for every word in the input text.
This work was done in the perspective of having a rapid and efficient ANLP tool.
Indeed, we have developed a POS tagger based on a hybrid approach containing two
main modules: a statistical module based on HMM, and the second one based on
decision trees. For the tag set we chose an elementary tag set to ensure accuracy and
rapidity which allowed us to develop a very rapid and accurate POS tagging system
with unknown words handling.
The evaluation of each module gives satisfying results (accuracy of 99% for HMM
and 97% for the tree-guessing model) and after the assembly of the two modules for
the complete POS tagger the results obtained are encouraging and the execution time
is very low.
As perspectives for the improvement of this work, we plan to enlarge the vocabu-
lary and increase the order of the hidden Markov model. It is also worth focusing on
the disparity of the POS tag set of the Arabic language and propose a standardized
version of both the tag set and the corresponding system.
References
1. Al Shamsi FG (2006) A hidden Markov model-based POS tagger for Arabic. In: Proceeding
of the 8th international conference on the statistical analysis of textual data, France, pp 31–42
2. Albared MO (2009) Arabic part of speech disambiguation, pp 517–532
208 K. Tnaji et al.
3. Attia MM (2005) Specifications of the Arabic Written Corpus produced within th NEMLAR
project
4. Atwell ES (2008) Development of tag sets for part-of-speech tagging
5. Atwell MS (2013) A standard tag set expounding traditional morphological features for Arabic
language part-of-speech tagging. Edinburgh University Press
6. Buckwalter Arabic Transliteration. (n.d.). https://www.qamus.org/transliteration.htm. 20 Oct
2020
7. Darwish K, Mubarak H, Abdelali A, Eldesouki M (2017) Arabic POS tagging: don’t abandon
feature engineering just yet. In: Proceedings of the third arabic natural language processing
workshop, pp 130–137. https://doi.org/10.18653/v1/W17-1316
8. Diab M, Hacioglu K, Jurafsky D (2004) Automatic tagging of Arabic text: from raw text
to base phrase chunks. In: Proceedings of HLT-NAACL 2004: short papers. association for
computational linguistics, pp 149–152
9. Dinçer BT, Karaoğlan B (2004) He effect of part-of-speech tagging on IR performance for
Turkish. In: Aykanat C, Dayar T, Körpeoğlu İ (eds.), Computer and Information Sciences—
ISCIS 2004, Springer, pp 771–778. https://doi.org/10.1007/978-3-540-30182-0_77
10. Habash NF (2009) Syntactic annotation in Columbia Arabic Treebank. In: 2nd International
Conference on Arabic Language Resources & Tools MEDAR. Cairo
11. Hammo B, Abu-Salem H, Lytinen SL, Evens M (2002) QARAB: A: question answering system
to support the Arabic language. In: Proceedings of the ACL-02 workshop on Computational
approaches to semitic languages. July 2002
12. Imad Zeroual AL (2017) Towards a standard Part of Speech tagset for the Arabic language. J
King Saud Univ Comput Inf Sci 171–178
13. Albared M, T-M O-S-A (2005) probabilistic Arabic part of speech tagger with unknown words
handling. J Theor Appl Inf Technol
14. Maamouri MA (2004) Developing an Arabic treebank: methods, guidelines, procedures, and
tools. In: Proceedings of the 20th international conference on computational linguistics
15. Rabiner L, Juang B (1986) An introduction to hidden Markov models. IEEE ASSP Mag 3(1):4–
16
16. Salameh S (2018) A review of part of speech tagger for Arabic Language. International Journal
of Computation and Applied Sciences IJOCAAS, 4, 4–5, June 2018. Darwish K, Mubarak H
(n.d.). Farasa: A New Fast and Accurate Arabic Word Segmenter. 5
17. Jaafar Y, Bouzoubaa K (2015) Arabic natural language processing from software engineering
to complex pipelines. In: Cicling 2015, Cairo, Egypt, April 2015
18. Zouaghi A, Merhbene L, Zrigui M (2012) Combination of information retrieval methods with
LESK algorithm for Arabic word sense disambiguation. Artif Intell Rev 38:257–269. https://
doi.org/10.1007/s10462-011-9249-3
RSEPUA: A Recommender System
for Early Predicting University
Admission
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 209
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_20
210 I. El Guabassi et al.
2 Related Work
Due to the rapidly growing interest in the field of education [4–6], there are several
research studies have been conducted on predicting university admission based on
several factors using supervised or unsupervised Machine Learning algorithms.
Xiaojun Wu & Jing Wu [7] conducted a study on predicting students’ selection
criteria in non-native language MBA admission based on Ridge regression, SVM,
Random forest, GBDT. Al Ghamdi et al. [8] applied three classification methods
which are linear regression, decision tree, and logistic regression model to auto-
matically predict postgraduate admission. Nandal [9] developed Student Admission
Predictor (SAP) based on Deep Learning techniques namely Decision Tree, Support
Vector Machine, Gaussian Naive Bayes, Linear Regression, Random Forest, and
Deep Neural Networks. Zhao et al. [10] explained their study based on a quantitative
Machine Learning approach to predict master students’ admission in professional
institutions. They used standard SVM, S3VM, and SVM+, as well as their dataset,
which is collected from Northeastern University’s MS in Computer Science (MSCS)
program.
In this section we will discuss the three main elements to achieve our purpose, these
elements are Recommender Systems, Algorithms used, and evaluation methods.
3.2 Algorithms
There are many algorithms for predictive modeling Machine Learning. In the next
sections, we will present the algorithms used to build predictive models.
Linear Regression [14] is the most important algorithm in the field of Machine
Learning, especially supervised learning. It is a way to model a relationship between
a dependent variable and one or more independent variables. It consists of finding a
regression line straight line through the points.
Decision Tree [15] is the most widely used classification and prediction technique. It
is a tree structure, where each internal node with outgoing edges indicates a condition
on an attribute, each branch is an outcome of the test, and each leaf terminal node
represents a class label.
Support Vector Regression (SVR) [16] is also a very popular Machine Learning
technique used in both classification and regression. It is similar to Linear Regression
with only a few minor differences. SVR allows defining how much error is acceptable
in our predictive model and will find an appropriate line to fit the data.
Mean Square Error (MSE) [18] is the arithmetic mean of the squares of the predic-
tions between the model and the observations. This is the value to be minimized in
the context of a single or multiple regressions.
Root Mean Square Error (RMSE) is a standard way to measure the error in model
evaluation studies. It is the square root of the mean of the square of all of the errors.
After briefly determining the settings and describing basic definitions, in the next
section, we will present the methodology used to develop our recommender System
RSEPUA.
RSEPUA … 213
Machine Learning
Algorithms
Dataset Data
Preparation
Modeling
Deployment
&
Evaluation
Integration
4 Proposed Methodology
Admission to graduate schools is the most difficult and important step. The major
problem students often fall into when applying to graduate schools is they are not
aware of admission requirements. The purpose of this research work is to provide
a recommender system for early predicting university admission based on Machine
Learning algorithms. The methodology used in this work is composed of six steps.
Figure 1 presents an overview of this methodology.
Dataset: The data used in this research work is collected from a second version of
the data set named “Graduate Admission 2” [19]. It is inspired by the University
of California in Los Angeles Graduate Dataset. This data is available publicly even
on Kaggle which offers free datasets for training and evaluation. It is comprised of
500 rows with 8 features. The dataset contains various features which are considered
important during the enrollment and admission for master’s programs.
Table 1 represents an overview of dataset features used for training and testing.
214 I. El Guabassi et al.
Modeling: This step involves the conception of different Machine Learning algo-
rithms (e.g., regression, classification, clustering, etc.) that can be used for predicting
university admission.
Evaluation: This step is a core part of building our Machine Learning model. There
are different metrics of evaluations that can be used. The evaluation metrics used in
this research work are Mean Square Error, Root Mean Square Error, and R-squared.
Deployment and Integration: It is all of the tasks that make our recommender
system RSEPUA available for use.
The recommender system proposed in this paper is articulated in two parts as shown
in Fig. 2. The first part is the User-based datamining that contains sample data.
The second part focuses on Machine Learning algorithms, that use different regres-
sion algorithms (i.e., Linear Regression, Decision Tree, Support Vector Regression,
and Random Forest Regression) to Build a predictive model in order to predict
students’ admission for higher education. It should be noted that the parameters used
in this study are GRE Score, TOEFL Score, University Rating, SOP, LOR, CGPA, and
Research Experience. After presenting the proposed recommender system RSEPUA,
in the next section we will present the discussion of the obtained results.
RSEPUA … 215
User Model
User-based datamining
Modeling Evaluation
GRE Score
Predictive Model TOEFL Score
University Rating
SOP
LOR
Prediction of Graduate
CGPA
Admission Research
We used the Pearson Correlation Coefficient (PCC) [21] to evaluate the linear corre-
lation between the variables in our dataset. Indeed, this coefficient allows us to reduce
the number of features and determine the relationship between different parameters
profile (i.e., GRE Score, TOEFL Score, University Rating, SOP, LOR, CGPA, and
Research Experience) and the chance of admission. Table 2 illustrates the results
obtained for the Pearson Correlation Coefficients. Figure 3 illustrates a correlation
216 I. El Guabassi et al.
map of the table values. The results obtained demonstrate a high correlation between
CGPA, GRE score, and TOEFL score and the chance of admission.
40
35
25
20
15
10
0
GRE Score University LOR Research
RaƟng
Variables
After identifying the most suitable Machine Learning algorithm for predicting univer-
sity admission, it is therefore evident to ask which parameters have the most predictive
power. Indeed, the parameters with high importance are the engines of the prediction
and their values have a strong and significant impact on the outcome values. For
that reason, after training and validating a random forest, the variable importance
(VIMP) is calculated based on Mean increase error.
The normalized variable importance measure for each variable is represented in
Fig. 3. As we can see, the most important parameter is the Cumulative Grade Point
Average (CGPA). Hence, it is clear that there is a very strong link between the CGPA
and the chance of admission.
In the 21st century, University education is becoming a key pillar of social and
economic life. In fact, it plays a major role not only in the educational process but
also in ensuring two important things which are the prosperous career and finan-
cial security. However, predicting university admission can be especially difficult
because the students are not aware of admission requirements. For that reason, the
218 I. El Guabassi et al.
main objective of this research work is to build a predictive model for early predicting
university admission. Thus, the contributions were threefold: The first was to apply
several Supervised Machine Learning algorithms (i.e., Linear Regression, Support
Vector Regression, Decision Tree Regression, and Random Forest Regression) on our
dataset. The second purpose was to compare and evaluate algorithms used to create a
predictive model based on various evaluation metrics. The last purpose was to deter-
mine the most important parameters that influence the chance of admission. The
experimental results showed that the Random Forest Regression is the most suitable
Machine Learning algorithm for predicting university admission. Also, the under-
graduate GPA (CGPA) is the most important parameter that influences the chance of
admission. The major directions for future work are: (i) Applying techniques such
as clustering and artificial neural networks to have better predicting. (ii) Utilizing
dataset with massive size and diverse features to tackle the issue of scalability. (iii)
Exploiting few hybrid feature selection algorithms.
References
1. Chui KT, Fung DCL, Lytras MD, Lam TM (2020) Predicting at-risk university students
in a virtual learning environment via a machine learning algorithm. Comput Hum Behav
107:105584
2. Qazdar A, Er-Raha B, Cherkaoui C, Mammass D (2019) A machine learning algorithm frame-
work for predicting students performance: a case study of baccalaureate students in Morocco.
Educ Inf Technol 24(6):3577–3589
3. El Guabassi I, Al Achhab M, Jellouli I, El Mohajir BE (2016) Recommender system for
ubiquitous learning based on decision tree. In: 2016 4th IEEE International Colloquium on
Information Science and Technology (CiSt), pp 535–540. IEEE
4. Guabassi IE, Achhab MA, Jellouli I, Mohajir BEE (2016) Towards adaptive ubiquitous learning
systems. Int J Knowl Learn 11(1):3–23
5. El Guabassi I, Al Achhab M, Jellouli I, Mohajir BEE (2018) Personalized ubiquitous learning
via an adaptive engine. Int J Emerg Technol Learn (iJET) 13(12):177–190
6. Bousalem Z, El Guabassi I, Cherti I (2018) Toward adaptive and reusable learning content
using XML dynamic labeling schemes and relational databases. In: International Conference
on Advanced Intelligent Systems for Sustainable Development, pp 787–799. Springer, Cham.
7. Wu X, Wu J (2019) Criteria evaluation and selection in non-native language MBA students
admission based on machine learning methods. J Ambient Intell Hum Comput 1–13
8. AlGhamdi A, Barsheed A, AlMshjary H, AlGhamdi H (2020) a machine learning approach
for graduate admission prediction. In Proceedings of the 2020 2nd international conference on
image, video and signal processing, pp 155–158
9. Nandal P (2020) Deep learning in diverse computing and network applications student
admission predictor using deep learning. SSRN 3562976
10. Zhao Y, Lackaye B, Dy JG, Brodley CE (2020) A quantitative machine learning approach to
master students admission for professional institutions. International Educational Data Mining
Society
11. Di Noia T, Mirizzi R, Ostuni VC, Romito D, Zanker M (2012) Linked open data to support
content-based recommender systems. In: Proceedings of the 8th international conference on
semantic systems, pp 1–8
12. Herlocker JL, Konstan JA, Terveen LG, Riedl JT (2004) Evaluating collaborative filtering
recommender systems. ACM Trans Inf Syst (TOIS) 22(1):5–53
RSEPUA … 219
1 Introduction
In the real world, with a remarkable growth within the universe of measured data
warehouse sizes, analyzing the data in order to extract useful information is becoming
a necessity and a rich subject for several research projects [1]. Many application
areas adopt machine learning methods in their systems namely telecommunication,
shopping platforms, restaurants, economy, marketing, tourist targets, and medicine.
Over the last two decades, machine learning has entered the e-learning space as well
[2–5]. Thus, several machine learning algorithms have been exploited by researchers
to predict hidden patterns from educational settings [6–8]. The early prediction of
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 221
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_21
222 I. El Guabassi et al.
students at risk for academic failure is a crucial step in the learning process. Thus,
it must be made as soon as possible during the period of learning. The prediction of
student performance is necessary for:
• Providing a good quality education for students as well as for tutors;
• Reducing the number of dropouts;
• Increasing the rate of school completion;
• Improving educational outcomes for all students.
However, the real and major problems are: How to identify the “weak” students
who will need additional help to improve their performance? Also, which the suit-
able machine learning algorithm (i.e. predictive model) for predicting students’
academic performance? This research work evaluates and compares the effectiveness
of different machine learning algorithms. While there are many regression algorithms
for creating predictive models, this work concentrates on seven of them, which are
(1) ANCOVA, (2) Logistic Regression, (3) Support Vector Regression, (4) Log-
linear Regression, (5) Decision Tree Regression, (6) Random Forest Regression, and
(7) Partial Least Squares Regression. The outline of the present paper is as follows:
Sect. 2 contains recent studies regarding the specified area. The dataset used is briefly
described in Sect. 3. Section 4 concentrates on our methodology. Section 5 presents
the results and discussion. Finally, Sect. 6 presents the main conclusions considering
some future research directions.
2 Related Literature
techniques which are the J48 Classifier and Naive Baye’s classification. Their results
show that HEDM outperforms the results obtained in EDM.
In summary, many existing studies have made significant results in educational
data mining. However, most of them use classification methods for predicting
students’ academic performance. Moreover, there was very little focus on inter-
actional and parental involvement features.
3 Dataset
The data used for this work’s experimentation is collected from a first version of
the dataset named “Students’ Academic Performance Dataset (xAPI-Edu-Data)”
[18, 19]. It is, therefore, an open-source dataset available publicly on the Kaggle
dataset repository for academic and research purposes. The primary source of this
dataset used is from Elaf Abu Amrieh, Thair Hamtini, and Ibrahim Aljarah. There are
three main categories of the dataset features named “Demographic characteristics”,
“Academic background characteristics”, and “Behavioral characteristics”. Table 1
contains an overview of dataset features used for training and testing. It contains
three fields: feature, description, and type. It should be noted that there are two
major feature types, named “Nominal” and “Numeric”.
4 Methodology
Increasingly, E-learning has become an important tool for teaching and learning
around the world. Further, Learners have the opportunity to switch to distance
learning in various scientific areas anytime and anywhere [9, 20]. It is therefore
evident that many researchers work on the various aspects of e-learning [10–12].
The identification of the “weak” students and the factors affecting students’ academic
performance is a crucial step for successful learning. Hence, in the present paper, our
aim is evaluating student’s academic performance and identifying the factors that
influence academic performance using supervised machine learning algorithms.
This research work focuses on the following steps:
• Applying several machine learning algorithms which are ANCOVA, Logistic
Regression, Support Vector Regression, Log-linear Regression, Decision Tree
Regression, Random Forest Regression, and Partial Least Squares Regression.
• Comparing and evaluating machine learning algorithms for identifying which are
most suitable by using several evaluation metrics which are Mean Square Error
(MSE), Root Mean Square Error (RMSE), and R-squared (R2 ).
Many experiments were conducted in seven major steps depending on the regres-
sion methods namely ANCOVA, Logistic Regression (Logit-R), Support Vector
Regression (SVR), Log-linear Regression (Log-LR), Decision Tree Regression
224 I. El Guabassi et al.
(DTR), Random Forest Regression (RFR), and Partial Least Squares Regression
(PLS-R). These regression methods were applied using the XLSTAT environment
[28]. In the following, the experimental result of each algorithm is presented.
4.1 Ancova
Support Vector Regression (SVR) [23] is a binary classification algorithm. Just like
the Logistic Regression. If we take the image above, we have two classes (e.g.,
suppose these are e-mails, and Spam mails are in red and non-spam emails are in
blue).
The table above (Table 4) shows the evaluation metrics obtained after evaluating
the Support Vector Regression Algorithm. Mean Square Error (MSE) is the arithmetic
mean of the squares of the predictions between the model and the observations. This
is the value to be minimized in the context of a single or multiple regressions.
The log-linear regression (Log-LR) [24] is a prediction model that applies when
the target variable Y is a counting variable (number of occurrence of events during
a lap of time). Table 5 contains the evaluation metrics obtained after applying the
Log-linear regression Algorithm.
Root Mean Square Error (RMSE) represents a standard way to measure the error
in the predictive model. It is the square root of the mean of the square of all of the
errors.
The decision trees (DTR) [25] are classification rules that base their decision on a
suite of tests associated with attributes, the tests being organized in a tree structure.
We evaluate our dataset using the Decision Tree Algorithm. Table 6 represents an
overview of the results of the evaluation.
Forecasting Students’ Academic Performance … 227
Partial Least Squares Regression (PLS-R) [27] is a flexible statistical technique appli-
cable to any form of data. It allows to model the relationships between inputs and
outputs, even when the inputs are correlated and noisy, the outputs multiple and the
inputs more numerous than the observations.
Our predictive model based on Decision Trees is evaluated using several evalu-
ation metrics such as Mean Square Error, Root Mean Square Error, and R-squared
(see Table 8).
228 I. El Guabassi et al.
The table below, therefore, represents summary results for the seven algorithms used
in this research work. The evaluation metrics used in this experiment are Mean Square
Error (MSE), Root Mean Square Error (RMSE), and R-squared (R2 ). It should be
noted that RMSE is just the square root of the MSE (Table 9).
After rigorously evaluating all the seven algorithms on the 480 students of our
dataset, we compare the performances in order to determine the most suitable predic-
tive model. According to the experimental results, it is clear that Log-linear Regres-
sion (Log-LR) provides better performance because it has a low MSE, low RMSE,
and high R2 score. On the other hand, we observed that Support Vector Regression
(SVR) isn’t suitable for predicting students’ academic performance because it has a
high MSE, high RMSE, and low R2 score.
Given the R2 = 73% of the variability of the dependent variable, Class is explained
by the 16 explanatory variables. The remainder of the variability is due to other
explanatory variables that have not been considered during the present experiment
research.
The chart below (see Fig. 1) indicates the predicted values versus the observed
values. Also, Confidence intervals for the mean allow the detection of potential
outliers.
The histogram below (see Fig. 2) represents the standardized residuals versus the
performance. It indicates that the residuals grow with the Performance. As we can
see in Fig. 2 the residuals bar chart allows to quickly showing the residuals that are
out of the range [2, 2].
Forecasting Students’ Academic Performance … 229
Pred(Performance) - Performance
5
3
Performance
0
-1 0 1 2 3 4 5
-1
Pred(Performance)
Standardized resudials/Performance
Obs457
Obs433
Obs409
Obs385
Obs361
Obs337
Obs313
Observations
Obs289
Obs265
Obs241
Obs217
Obs193
Obs169
Obs145
Obs121
Obs97
Obs73
Obs49
Obs25
Obs1
-4 -2 0 2 4
Standardized resudials
In recent years, the world’s population is increasingly demanding to predict the future
with certainty, predicting the right information in any area is becoming a necessity.
One of the ways to predict the future with certainty is to determine the possible
futures. In this sense, machine learning is a way to analyze huge datasets in order to
make strong predictions or decisions. In this research work, our main purpose was
to compare several machine learning algorithms for predicting student’s academic
performance. Certainly, this research work has some limitations. That’s why the
major directions for future work could focus on the following: Firstly, applying
techniques such as clustering and artificial neural networks to have better predicting.
Secondly, utilizing dataset with massive size and diverse features to tackle the issue
of scalability. The final area that can be improved is exploiting few hybrid feature
selection algorithms.
References
13. Bravo-Agapito J, Romero SJ, Pamplona S (2020) Early prediction of undergraduate student’s
academic performance in completely online learning: a five-year study. Comput Human Behav
106595
14. Gray CC, Perkins D (2019) Utilizing early engagement and machine learning to predict student
outcomes. Comput Educ 131:22–32
15. Hamsa H, Indiradevi S, Kizhakkethottam JJ (2016) Student academic performance prediction
model using decision tree and fuzzy genetic algorithm. Proc Technol 25:326–332
16. Hussain M, Zhu W, Zhang W, Abidi SMR, Ali S (2019) Using machine learning to predict
student difficulties from learning session data. Artif Intell Rev 52(1):381–407
17. Karthikeyan VG, Thangaraj P, Karthik S (2020) Towards developing hybrid educational data
mining model (HEDM) for efficient and accurate student performance evaluation. Soft Comput
1–11
18. Amrieh EA, Hamtini T, Aljarah I (2016) Mining educational data to predict student’s academic
performance using ensemble methods. Int J Database Theory Appl 9(8):119–136
19. Amrieh EA, Hamtini T, Aljarah I (2015, November) Preprocessing and analyzing educational
data set using X-API for improving student’s performance. In: 2015 IEEE Jordan conference
on applied electrical engineering and computing technologies (AEECT), pp. 1–5. IEEE
20. El Guabassi I, Bousalem Z, Al Achhab M, Jellouli I, Mohajir BEE (2018) Personalized adaptive
content system for context-aware ubiquitous learning. Proc Comput Sci 127:444-453
21. Rutherford A (2001) Introducing ANOVA and ANCOVA: a GLM approach Sage
22. Kleinbaum DG, Dietz K, Gail M, Klein M, Klein M (2002) Logistic regression. Springer, New
York
23. Smola AJ, Schölkopf B (2004) A tutorial on support vector regression. Stat Comput 14(3):199–
222
24. Heien DM (1968) A note on log-linear regression. J Am Stat Assoc 63(323):1034–1038
25. Safavian SR, Landgrebe D (1991) A survey of decision tree classifier methodology. IEEE Trans
Syst Man Cybern 21(3):660–674
26. Liaw A, Wiener M (2002) Classification and regression by randomForest. R news 2(3):18–22
27. Geladi P, Kowalski BR (1986) Partial least-squares regression: a tutorial. Anal Chim Acta
185:1–17
28. Addinsoft X (2015) Data analysis and statistics with MS Excel. Addinsoft, NY, USA. xlstat
available at https://www.xlstat.com/en/home
Computational Analysis of Human
Navigation Trajectories in a Spatial
Memory Locomotor Task
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 233
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_22
234 I. Annaki et al.
1 Introduction
Among the most fundamental human tasks is to navigate from an initial position to
a destination. This process is performed several times a day by each of us. In fact,
navigation is characterized as organized and goal-oriented environmental movements
[3]. This includes the physical act of action and the cognitive dimensions of decision-
making and an itinerary follow-up. These two navigational elements are known as
locomotion, spatial orientation, and navigation [4].
To assess individuals’ efficiency at the highest level of locomotive control, i.e.,
navigation [5]. Neuropsychologists use a locomotor version of a table memory task
called the Corsi Block Tapping task (CBT) [15]. They use the Walking Corsi Task
(WCT) and its variants [2] and [6], which is closely related to navigation. The
performance data resulting from the WCT has been analyzed previously in studies
concerning the pathological deficits in adults and children, but the present work brings
new methods for a Kinematic analysis of the trajectories and mental manipulation
of viewpoint changes using [7].
The previous analysis of trajectories that a human may take to reach a distant
position and turn in a given direction revealed that the formation of trajectories
during goal-oriented locomotion in humans is, in practice, stereometric and kinemat-
ically stereotypical [8, 9]. Furthermore, robotics engineers and researchers are more
concerned with implementing pathfinding algorithms that mimic human behavior
and can also contribute to optimizations based on various parameters, such as time
and space [10].
In neurosciences however classical statistical approaches have been used to test
differences in behavior and navigation strategies between individuals and groups.
This has prompted our attempt to develop more effective methods to help clinicians
in sorting out normal subjects and patients or for the study of the development of
sensory-motor and cognitive abilities [11].
From this perspective, the present paper aims to explore the extracted data
outcomes of the WCT test from a kinematic point of view with data analytics
using the potential of python and computer simulations developed with Unity. It
represents, a new effort toward integrating in-depth artificial intelligence, computer
vision, and deep learning approaches, and stepping into computer simulations by
taking advantage of the capacity of virtual reality. It is a specific contribution to
the wider developing field using information contained in experimental data and
combining information from different sources and different nature and coping with
a large amount of data (Big-Data Problem) [12]. Also, the effort here to generate a
synthetic data analysis based on real samples provided from clinical trials using data
augmentation models [13]. It also is applied to the virtual environment for simulations
or enhancements of the neuropsychological assessment [14].
Computational Analysis of Human Navigation Trajectories … 235
The Corsi visuospatial memory protocol was used in a series of studies using different
technologies. These protocols were all inspired by the Corsi Table Block Tapping test
(CBT). A locomotor version of this table test was developed using simple targets.
Then two new technologies were developed the “Magic Carpet” and the “Virtual
Carpet”. They allowed more complex measurements of the behavior and more flexi-
bility in the presentation of the targets and kinematic measurements of the trajectories
and behavior of the subjects.
output—specifically, the number of correct tiles achieved in the right order. The
series to be memorized is presented by illuminating the tiles in series consecutively
at one tile per 2 s (each tile lights up for 2 s and then shuts off before the next tile
is lit). At the end of the experiment, the auditory alarm signals to the participant
that the pattern detected while walking must be repeated. This Magic Carpet Corsi
protocol was used for several studies on normal and pathological adults and children
and the processing of correct target acquisitions, delays, and several kinds of errors
were computed.
The batch contains a calibration file that sets the coordinates of the targets relative to
the current subject under focus, and each corresponding folder contains the derived
data sets relative to the trajectory and locomotion.
Computational Analysis of Human Navigation Trajectories … 237
2.3 Participants
The data collected were provided without further detail on the subjects, from gender
and age to actual status so that the analysis would not have any a priori bias. However,
this sample was subject to the usual protocols, and participants were made aware that
their data would be analyzed, and they could have access to the results. Besides, the
details have been obscured to make the data more accessible and to examine it
extensively without any background restrictions.
Using Matplotlib, a robust library for developing static, interactive, and immersive
Python visualizations and based on our theoretical claims. We visually repeated the
sequence on which our findings were based. The reproduction of the trajectory and
the targets with animation allowed us to understand the kinematic behavior of the
subjects during each session (Fig. 1).
This plot represents the experimental array seen from above each target is shown
as a square. The departure point is on the right. The subject had to go to two targets
and return. Head trajectory is shown, and the blue circles indicate that the subject
stayed successfully on both targets.
Several computations have been carried out on the data to establish a firm inter-
pretation and find criteria to help us draw statements about the various subjects.
We developed an algorithm to detect the various targets visited. This was done by
computing the head positions during the trajectory and from this set of numbers
calculate a walking velocity along the trajectory (tangential velocity). To obtain the
information that the subject had reached a target we compare the position data with
a target calibration field in which the X and Y components of the nine targets were
given. This allowed us to also compute the duration in seconds during which the
subject had stopped on the target. This file implements nine target positions, as each
dataset in the user folder reflects a series of scenarios.
Modern techniques sallow computational approaches of these data [19] and [20].
Using “Pandas” a data analysis and manipulation tool library, in particular the concept
of DataFrame, each file is stored as a data frame to avoid any conflict between the
file types. Each target is also flagged to be able to settle on the series assigned to
each subject. The crucial step in the algorithm is to compare the distance between
the positions of the subject on the trajectory and the location of the target extracted
from the calibration file, to be able to determine whether the target was visited by
the subject during the session, and to track the kinematic behavior until the subject is
close to the target, or until the entire session is over. This potentially can also indicate
if the subject missed a target.
Using all these data we could classify the different subjects according to the values
of velocity, and time on various targets. We also distinguished the behavior when the
trajectory included only 2 targets or more up to Nine which is a much more difficult
task for instance in the Table Block test a span of 7 is generally the maximum that
most normal subjects can memorize.
3 Results
We took as a criterion that a subject had visited a target if his position was within a
confidence circle of 30 × 30 cm, around the value given by the calibration file. A
velocity slowdown was also taken as a criterion of target reaching. A first numerical
calculation has allowed us to identify a classification of experiments according to
Computational Analysis of Human Navigation Trajectories … 239
the number of targets visited by the subjects. The result is a classification of the
data according to the sequence’s length (number of visited targets). It indicates four
classes. These classes referred to reflect the distribution of subjects according to the
series correctly replicated by the subject. Specifically, we find classes in which the
participant repeats situations of series lengths of 2, 3, 4, and 5. As a side effect, more
classes were recognized, however, the number of subjects in these classes does not
allow us to extract significant information.
We aggregate the subject’s data to retrieve their average speed along the entire trajec-
tory and the time, they remain on a target before leaving the confidence area as
mentioned in the target detection algorithm. The global analysis allows us to evaluate
the performance of a subject throughout the whole process and determine aggregates
related to speed. These aggregates are a comparison threshold allowing us to classify
the subjects into 3 categories:
– The fastest subjects.
– The moderately fast subjects.
– The slowest subjects.
The aggregate of the global average time that the subject spends within a
target allows us to compare the coherence between speed and time and allows us
to identify anomalies and classify the subjects into 2 categories:
– Subjects spending important time within a target.
– Subjects moderately spending below or closer time to the global average of time
spent within a target (Fig. 2).
The same approach is used for general analysis. For each class, we aggregate the data
of the subject to extract its average speed and the time spent in the target. Every class
helps us to calculate the class’s speed and time spent within the target aggregates to
extract the same categories as in the general classification.
The detailed analysis allows us to identify subject participation and attendance at
the level of each class and to refine the overall analysis carried out (Fig. 3).
240 I. Annaki et al.
Fig. 2 General Combination Chart representation of average speed as a red line chart and time
spent on the target as a blue-sky bar chart. The green reference line represents the overall average
speed (All subjects), the red one is the average speed of slow subjects (the qualification of slow
subjects is based on the subject’s speed below the overall average speed) and the blue line identifies
the average speed of fast subjects (as cited for the slow subjects, the fast ones are those who got a
speed over the overall average speed)
Fig. 3 Example: Class 3 combination chart representation of average speed as a red line chart
and time spent on the target as a blue-sky bar chart. The green reference line represents the class
average speed (All subjects attending the scenarios with 3 targets visited), the red is the average
speed of slow subjects (the qualification of slow subjects is based on the subject’s speed below the
class average speed) and the blue identifies the average speed of fast subjects (as cited for the slow
subjects, the fast ones are those who got a speed over the class average speed)
Computational Analysis of Human Navigation Trajectories … 241
4 Discussion
The purpose of this research was to identify the different behaviors of subjects who
performed the Walking Corsi Protocol on the Virtual Carpet set up using a kinematic
approach instead of basing the research on pre-established scoring strategies in the
literature. Commonly, scoring is based on the maximum length of the series that the
subject can achieve correctly. Here we characterized the behavior and performance
used the subject using average speed to complete the sequences and the time spent
on the targets before heading to another goal. Moreover, we were able to identify
some incoherence between the subject’s pace and the spent time. Further studies are
needed to generalize this approach and analyze its efficiency and accuracy.
This is a very global approach that will complement other data analysis of the
behavior, but it can be very useful to provide criteria that give us insights about
the subjects who can repeat the sequence easily, and other subjects who find it
challenging, particularly when the sequence is getting longer or requires mental
rotations, etc. None of the existing publications on these tests has brought this tool
and it is worth trying to use it in young children or with patients or elderly persons
and eventually it may be very useful to measure quantitatively learning or functional
recuperation during rehabilitation.
5 Conclusion
The current paper provides the first attempt at characterizing the behavior of human
subjects in a complex visuospatial task used to test memory and navigation capabili-
ties. The aim is to integrate emerging technology into the psychometric methods that
will give neuropsychologists greater insight into new technology, such as big data,
machine learning, and computer vision [20]. Knowing that psychometric clinical
trials require a significant amount of time and planning, it can take months to years
to collect data. Then data science can offer new approaches to generate synthetic
data based on the samples extracted from the psychometric tests. Neuropsycholog-
ical assessments are valuable instruments for telling cognitive impairment diagnosis.
However, decoding neuropsychological assessments requires experts and is often
time-consuming. To streamline the use of psychometric evaluations in clinical envi-
ronments, machine learning algorithms are being built and tested using multi-center
neuropsychological data [21].
Acknowledgments We thank Dr. Bernard Cohen, Paris France, to let us use the preliminary test
data obtained in cooperation with him to test the present method to evaluate its validity for the use
of patient assessments.
242 I. Annaki et al.
References:
19. Berthoz A, Zaou, M (2015) New paradigms and tests for evaluating and remediating visuospa-
tial deficits in children. Dev Med Child Neurol 57(Suppl 2):15–20. https://doi.org/10.1111/
dmcn.12690
20. Elgendy N, Elragal A (2016) Big data analytics in support of the decision-making process.
Procedia Comput Sci 100:1071–1084. https://doi.org/10.1016/j.procs.2016.09.251
21. Kang MJ, Kim SY, Na DL et al. (2019) Prediction of cognitive impairment via deep learning
trained with multi-center neuropsychological test data. BMC Med Inform Decis Mak 19:231.
https://doi.org/10.1186/s12911-019-0974-x.
An Electrocardiogram Data
Compression-Decompression Technique
Based on the Integration Filtering
of Peaks’ Amplitude Signal
Skander Bensegueni
1 Introduction
The electrocardiogram (ECG) data are the result of the electric potential recording of
the myocardium. It’s used to monitor the heart’s health. The need for long duration
recording or permanent monitoring of the electrocardiogram signal is gaining more
and more attention for its ability to record pathological events appearing suddenly.
The compression of ECG signal is very important in health monitoring, medical diag-
nosis aid and decision systems for patients with cardiovascular diseases. By reducing
the size of the cardiac activity record, it will be easy to store and transmit data, which
allow the analysis and processing of this data distantly and by automatic tools.
The construction of an efficient ECG telemonitoring system can be achieved
by reducing the amount of transmitted data and minimizing the computational
S. Bensegueni (B)
National Polytechnic School, BP 75, A, Nouvelle Ville RP, Constantine, Algeria
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 245
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_23
246 S. Bensegueni
complexity [1]. In the paper of Polanía et al. [2], a reducing dimensionality tech-
nique was proposed by exploiting the representational power of restricted Boltz-
mann machines (RBMs) to model the probability distribution of the sparsity pattern
of transmitted ECG signals. Several ECG data compression techniques have also
been proposed in the literature. Techniques using wavelets give good performances;
such as the dimensionality reduction approach applied in [3] by selecting elementary
components from redundant wavelet dictionaries via a greedy strategy. The statis-
tical methods also were used to compress ECG data as in [4] by varying discrete
cosine transform thresholds applied on different principal directions of the ECG
data. A sparse representation-based compression algorithm was proposed by Grossi
et al. [5] through exploiting the repetition of common patterns of ECG segments as
natural basis.
The proposed compression algorithm is based on the application of transformation
of the ECG signal into peaks’ amplitudes signal which will contain a reduced number
of non-zeros amplitudes to be transmitted. This paper is divided into several sections;
starting from the development of the new theoretical basics as: the slop change
coefficients and the integrating filter, going through the application of these concepts
to compress and decompress the electrocardiogram signal, and finishing by results
and conclusions.
2 Theory
We calculate the slop change (SC) coefficients by using (1) in order to decom-
pose the signal x(t) into coefficients characterized by their sensitivity to steepness
changes where lower slopes have coefficients’ absolute values lower than those of
sharper slopes. The SC coefficients are constructed by calculating the integral of the
combination of two scale signals with opposite signs [6].
k k+1/2
SC(k) = − ∫ x(t − k)dt + ∫ x(t − k)dt (1)
k−1/2 k
l +∞
y(k) = − ∫ g(l − 2k)x(t − l)dt
l−1/2 l=−∞
(4)
l+1/2 +∞
+ ∫ g(l − 2k)x(t − l)dt
l k=−∞
The resulting discrete signal y(k) will detect, by exploiting the proprieties of
high-pass filter, the min.-max. (resp. max-min.) couples of SC coefficients and will
interpolate the samples by 2. The maxima (resp. minima) of the signal x(k) are
represented by two successive positive (resp. negative) y(k) samples (Fig. 1.c).
In a final step, we calculate the peaks’ amplitude (PA) of the signal, by the
following procedure:
1) If y(k) and y(k + 1) have the same sign, then: P A(k + 1) = 1
2) If y(k) and y(k + 1) have opposite signs, then: P A(k + 1) = 0.
(a) (b)
(c) (d)
Fig. 1 Changing in slopes detection: (a) double triangular signal, (b) SC coefficients, (c) high-pass
interpolation filter and (d) PA signal
248 S. Bensegueni
The PA signal will detect all the samples positions of the changes of slopes’
directions (i.e. position of optima values of the signal).
The purpose of the integrator filter is to merge the amplitudes of separate non-zero
samples by connecting them with regressive or progressive lines. In practice, this
integrating filter is calculated using a moving window defined by the following
equation:
1 N −1
y(k) = x(k − l) (5)
N l=0
3 Compression Stage
The slop change coefficients of the ECG signal are generated by applying (1), in order
to detect all the changes of the directions of this signal. The resulting coefficients
will represent peaks by two successive coefficients with the same sign, and the
other samples of the signal will be represented by a sequence of coefficients having
alternating opposite signs.
An Electrocardiogram Data Compression-Decompression Technique … 249
Fig. 2 Detecting of
changing in direction of the
ECG signal. All signal peaks
are represented by two
successive coefficients with
the same sign
After calculating the SC coefficients, the time scale of this resulting signal, will be
reduced to half the time size of the original signal. Then, to recover the initial signal
size, we apply a high-pass interpolation filter g on SC coefficients by using (4).
The resulting signal will detect the min.-max. (resp. max.-min.) couples of SC
coefficients and will interpolate the samples by 2 (Fig. 1.d). The local optima values
of the ECG signal will be represented by two successive positive (resp. negative)
samples.
The final step of the proposed compression method is to generate a peaks’ amplitudes
(PA) signal which produces a signal with amplitudes equal to those of the ECG signal
for the filtered SC min.-max. and max.-min. couples (i.e. at the peak positions of the
ECG signal) and zero amplitudes for the other samples.
Figure 3.b shows the PA of an electrocardiogram signal. We note that the samples
of the ECG signal with changes of slopes’ directions are detected and replaced at
peaks positions by pulses having the same amplitudes values. The remaining samples
are replaced by zeros.
The non-zeros PA coefficients coded on 32 bits (simple precision) and the null
coefficients coded on one 1 bit, will be stocked or transmitted to the decompression
bloc as shown in Fig. 4.
4 Decompression Stage
In this stage, the received coefficients will be used to reconstruct the ECG signal
by using an integrator filter in order to merge the amplitudes of separate non-zero
samples and connecting them with regressive or progressive lines by using (5) and
250 S. Bensegueni
replacing the x(k) samples by the received PA coefficients. But before that, we should
convert the one-bit zero samples to 32 bits (simple precision) samples.
In general, the choice of the width of the integrator filter window should be
approximately the same as the smallest ECG characteristic wave. Actually, the width
of the window is determined empirically, and for a sample rate of 360 Hz, the window
is 7 samples width (approximately equal to 19 ms).
where ecg is the original signal and ecgr is the reconstructed signal from the
decompression part;
Compression Ratio. The (CR) is one of the most used measures for evaluating
compression/decompression techniques performances. It is defined by:
Quality Score. The (QS) is used to evaluate the trade-off between the compression
ratio and the PRD by using the following ratio:
CR
QS = (8)
P RD
An Electrocardiogram Data Compression-Decompression Technique … 251
(a)
(b)
(c)
Fig. 3 Changing in slopes detection: a the ECG signal, b the PA signal and c in red, the
reconstruction of the ECG signal by using the high-pass interpolation filter
Computational Speed. The (CS) is defined as the speed at which the computer
execute the code and process data. It’s calculated by:
Decompression
Slop Change steps
Original ECG
Coefficients of ECG
signal
Signal
7 samples wide
integrator filter
Signal Time Scale
Recovery
Reconstructed
Generation of Null ECG signal
Compression Samples
steps
The previously defined criteria were applied on data from the MIT-BIH database,
using the Matlab environment installed on a SSD drive of an i5 3rd generation laptop
with 2.60 GHz 2 cores and 4 threads processor, and equipped with a 8 GB of DDR3.
The proposed method was applied on all the 30 min ECG records of the MIT-BIH
database, starting by calculating the slop change coefficients of these ECG signals
and by recovering their time scales. In a final step, the generation of null samples is
achieved in order to reduce the dimension of the records. Table 1 shows the evaluation
criteria values of the compression/decompression method of the 48 MIT-BIH records.
In Fig. 5, we can see the subplots of the reconstructed and the original signal of
segments of records: 100, 221 and 223. These figures show that the reconstructed
and the original signals are almost identical even in segments with sudden changes
in slope.From the previous results, we can verify that the mean value of the PRD
is equal to 0.62. This value shows that the reconstructed signals are acceptable and
the existent distortions are negligible and unimportant for many biomedical uses,
especially if we compare this value with the CR mean value equal to 17.83 which
represents a good ratio of signal compression.
An Electrocardiogram Data Compression-Decompression Technique … 253
Table 1 (continued)
Records PRD CR QS CS (MIPS)
217 0.88 23.66 26.89 2.04
219 0.56 19.10 34.11 1.94
220 0.60 20.98 34.97 2.01
221 0.65 17.98 27.66 2.09
222 0.42 15.82 37.66 2.03
223 0.78 21.92 28.10 2.10
228 0.59 16.00 27.12 1.28
230 0.73 19.40 26.57 1.26
231 0.68 18.48 27.18 1.39
232 0.47 14.22 30.26 1.36
233 0.76 19.24 25.32 1.39
234 0.55 15.34 27.89 1.30
(a) (b)
(c)
Fig. 5 Original and reconstructed signals of MIT-BIH records segments: a 100, b 221 and c 223
An Electrocardiogram Data Compression-Decompression Technique … 255
5 Conclusion
References
1. Lalos AS, Alonso L, Verikoukis C (2014) Model based compressed sensing reconstruction
algorithms for ECG telemonitoring in WBANs. Digital Signal Process 35:105–116
2. Polanía LF, Plaza RI (2018) Compressed sensing ECG using restricted Boltzmann machines.
Biomed Signal Process Control 45:237–245
3. Rebollo-Neiraa L, Cerná D (2019) Wavelet based dictionaries for dimensionality reduction of
ECG signals. Biomed Signal Process Control 54:1–9
4. Bensegueni S, Bennia A (2016) ECG signal compression using a sinusoidal transformation of
principal components. Int J Softw Eng Appl 10:59–68
5. Grossi G, Lanzarotti R, Lin J (2015) High-rate compression of ECG signals by an accuracy-
driven sparsity model relying on natural basis. Digital Signal Process 45:96–106
6. Bensegueni S A new method for electrocardiogram features extraction using slope change
coefficients (under review)
7. Moody GB (1992) ECG database applications guide. Harvard University
8. Tiwari A, Falk TH (2019) Lossless electrocardiogram signal compression: a review of existing
methods. Biomed Signal Process Control 51:338–346
Internet of Things, Embedded Systems,
Blockchain and Security
Narrowband IoT Evolution Towards 5G
Massive MTC: Complete Performance
Evaluation
1 Introduction
The Internet of Things (IoT) market for the 5G ecosystem is based on innovative tech-
nologies that efficiently support massive Machine-Type Communications (mMTC).
The 5G system is a next-generation wireless network that is based on New Radio
(NR) technology and is also designed to provide connectivity to a massive number
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 259
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_24
260 A. Abou El Hassan et al.
2 NB-IoT Overview
To ensure deeper coverage, the bandwidth occupied by the NB-IoT carrier is 180 kHz
corresponding to an one Physical Resource Block (PRB) of 12 subcarriers in an LTE
system [4]. There are three operation modes to deploy NB-IoT: as a stand-alone
Narrowband IoT Evolution Towards 5G Massive MTC … 261
carrier, in guard-band of an LTE carrier and in-band within an LTE carrier, which
offers operating mode flexibility of NB-IoT [6].
In order to coexist with LTE system, NB-IoT uses orthogonal frequency divi-
sion multiple access (OFDMA) in downlink with the identical subcarrier spacing
of 15 kHz and frame structure as LTE [4]. Whereas NB-IoT uses in uplink single-
carrier frequency division multiple access (SC-FDMA) and two transmission modes
which are the multi-tone and single-tone transmissions to ensure both high capacity
and maximum coverage for NB-IoT device with a single antenna [4, 6]. Multi-tone
transmission uses the same 15 kHz subcarrier spacing and 0.5 ms slot duration as
LTE. While single-tone transmission supports two numerologies which use 15 kHz
and 3.75 kHz subcarrier spacings with 0.5 ms and 2 ms slot durations respectively
[7]. Also, the restricted QPSK and BPSK modulation schemes are used in downlink
and uplink [6].
Two device categories Cat-NB1 and Cat-NB2 are defined by NB-IoT which corre-
spond to the device categories introduced in Rel-13 and Rel-14 respectively. The
maximum transport block size (TBS) supported in uplink by Cat-NB1 is only 1000
bits compared to 2536 bits for Cat-NB2. For downlink, the maximum TBS supported
by Cat-NB1 is only 680 bits compared to 2536 bits for Cat-NB2 [3].
The signals and channels used in downlink (DL) are as follows: Narrowband Primary
Synchronization Signal (NPSS), Narrowband Secondary Synchronization Signal
(NSSS), Narrowband Reference Signal (NRS), Narrowband Physical Broadcast
Channel (NPBCH), Narrowband Physical Downlink Shared Channel (NPDSCH)
and Narrowband Physical Downlink Control Channel (NPDCCH) [4].
Since NB-IoT uses only one PRB in a slot, the downlink signals and channels
are therefore multiplexed only in the time domain. The NB-IoT downlink frame is
structured as follows: subframe #0 of every frame carries NPBCH, subframe #1 to
#4, and subframe #6 to #8 of every frame carry NPDCCH or NPDSCH. NPSS is
carried by subframe #5 of every frame, whereas NSSS is carried by subframe #9
of every even numbered frame. While subframe #9 of every odd numbered frame
carries NPDCCH or NPDSCH [7].
In the uplink (UL), only one signal and two channels are used: Demodu-
lation Reference Signal (DMRS), Narrowband Physical Uplink Shared Channel
(NPUSCH) and Narrowband Physical Random Access Channel (NPRACH) [4].
Two formats are used for NPUSCH which are: Format 1 (F1) and Format 2 (F2).
NPUSCH F1 is used by user equipment (UE) to carry uplink user’s data to the
evolved Node B (eNB), and it supports both single-tone and multi-tone transmis-
sions [5, 7]. Whereas, NPUSCH F2 is used to carry uplink control information (UCI),
such as Hybrid Automated Repeat reQuest-Acknowledgement (HARQ-ACK), and
it supports only single-tone transmission [5, 7].
262 A. Abou El Hassan et al.
For cell access, the UE must first synchronize with the eNB using NPSS and
NSSS signals to achieve time and frequency synchronization with the network and
cell identification. Then, it receives Narrowband Master Information Block (MIB-
NB) and System Information Block 1 (SIB1-NB) carried by NPBCH and NPDSCH
respectively from eNB to access the system. Thereafter the UE starts random access
procedure by using single-tone transmission of NPRACH with 3.75 kHz subcar-
rier spacing, and waits for the scheduling uplink grant on NPDCCH which is used
to convey Downlink Control Information (DCI) for uplink, downlink and paging
scheduling [4, 5].
In Rel-15, 3GPP has defined five targets for 5G mMTC corresponding to the following
performances: coverage defined by the maximum coupling loss (MCL), throughput,
latency, battery life and connection density [8]. In this section, a complete evaluation
of these five performances is presented to verify whether the 5G mMTC targets are
fulfilled by NB-IoT. In addition, the enhancements of performance provided by the
recent 3GPP releases are also discussed.
3.1 Coverage
The MCL is a common measure to define the level of coverage a system can support.
It is depending on the maximum transmitter power (PTX ), the required signal-to-
interference-and-noise ratio (SINR), the receiver noise figure (NF) and the signal
bandwidth (BW) [9]:
Based on the simulation assumptions given in Table 1 according to [10] and using
(1) to calculate MCL, Table 2 provides the NB-IoT channels coverage to achieve the
MCL of 164 dB which corresponds to the 5G mMTC coverage requirement to be
supported [8].
Table 2 also shows the required acquisition time and block error rate (BLER)
associated with each channel to achieve the targeted MCL of 164 dB.
From Table 2, we note that to reach the MCL of 164 dB at the appropriate BLER, it
is necessary to use the time repetition technique of the simulated channels. However,
NPUSCH F1 is considered as the limiting channel, since it uses the longest acquisition
time to achieve the MCL of 164 dB. Whereas NPDCCH must be configured with
512 repetitions to achieve the targeted BLER of 1%, knowing that the maximum
configurable repetition number for NPDCCH is 2048 [5]. Therefore, NB-IoT can
support device operations in extreme coverage.
3.2 Throughput
The downlink and uplink throughputs are obtained according to the NPDSCH and
NPUSCH F1 scheduling cycles respectively and based on the acquisition times shown
in Table 2.
264 A. Abou El Hassan et al.
Figure 1 depicts the NPDSCH scheduling cycle according to [10], where the
NPDCCH user-specific search space is configured with a maximum repetition factor
Rmax of 512 and a relative starting subframe periodicity G of 4.
Based on BLER and TBS given in Table 2 and using an overhead (OH) of 5
bytes corresponding to the radio protocol stack, a MAC-layer throughput (THP) in
downlink of 281 bps is achieved with the following formula:
(1 − B L E R)(T B S − O H )
THP = (2)
N P DCC H Period
The NPUSCH F1 scheduling cycle depicted in Fig. 2 corresponds to scheduling
of NPUSCH F1 transmission once every fourth scheduling cycle according to [10],
which ensures a MAC-layer THP in uplink of 281 bps according to the formula (2)
using the BLER and TBS given in Table 2 and OH of 5 bytes.
As part of 3GPP Rel-15, 5G mMTC requires that downlink and uplink throughputs
supported at the MCL of 164 dB must be at least 160 bps [8]. As can be seen,
the MAC-layer throughput meets the 5G mMTC requirement in both downlink and
uplink. It should be noted that the BLER targets associated with each channel require
the acquisition times shown in Table 2. Therefore, by choosing to operate with
NPDSCH and NPUSCH F1 BLER levels above 10% and the use of new Cat-NB2
device introduced in 3GPP Rel-14 which supports a larger TBS in both downlink
and uplink with two simultaneous HARQ processes [5], should further enhance the
throughput levels.
Narrowband IoT Evolution Towards 5G Massive MTC … 265
3.3 Latency
The latency should be evaluated for two cases: the first case is the use of the Radio
Resource Control (RRC) Resume procedure depicted in Fig. 3, while the second
case consists in use of new Early Data Transmission (EDT) procedure that has been
introduced in Rel-15 and depicted in Fig. 4. The EDT procedure allows the device to
terminate the transmission of small data packets earlier in RRC-idle mode without
switching to RRC-connected mode.
The latency evaluation is based on the same radio related assumptions and the
system model given in Table 1 by using the data and signaling flows corresponding
to RRC Resume and EDT procedures. Table 3 shows the used packet sizes and the
evaluated latency at the 164 dB MCL according to [10].
When the RRC Resume procedure is used, we obtain a latency of 9 s, compared
to 5.8 s when the EDT procedure is used. As can be seen the 5G mMTC target of
10 s latency at the MCL of 164 dB defined in 3GPP Rel-15 [8] is met, for both the
RRC Resume and EDT procedure. However, the best latency of 5.8 s is obtained by
using the EDT procedure, through multiplexing the user’s data with Message 3 on
NPUSCH F1 channel as shown in Fig. 4.
The RRC resume procedure is used in battery life evaluation instead of the EDT
procedure, since EDT procedure does not support uplink TBS larger than 1000 bits
which requires long transmission times. The packet flow used to evaluate battery life
is the same as shown in Fig. 3 where DL data corresponds to application acknowl-
edgment regarding receipt of UL report by the eNB. Four levels of device power
consumption are defined, including transmission, reception, Idle-Light sleep corre-
sponding to device in RRC-Idle mode or RRC-Connected mode but not actively
receiving or transmitting, whereas Idle-Deep sleep corresponds to power saving
mode. The device power consumption for each level and assumed traffic model
according to Rel-14 scenario are given in Table 4. In addition, an Active Timer of
20 s is included after connection release where the device is in Idle-Light sleep,
to monitor the control channel NPDCCH according to [11]. The Rayleigh fading
Extended Typical Urban (ETU) channel model with a Doppler shift of 1Hz is used
to evaluate the battery life according to [11].
The battery life in years is calculated using the following formula according to
[9]:
Table 4. Assumptions of
Message format
NB-IoT battery life evaluation
UL report 200 bytes
DL Application Acknowledgment 20 bytes
Report periodicity Once every 24 h
Device power consumption
Transmission power (PTx ) 500 mW
Reception power (PRx ) 80 mW
Idle - Light sleep power (PILS ) 3 mW
Idle - Deep sleep power (PIDS ) 0.015 mW
Knowing that Eday is the device’s energy consumed per day, expressed in Joule and
calculated as follows:
E day = (PT x × TT x + PRx × TRx + PI L S × TI L S ) × Nr ep + (PI DS × 3600 × 24)
(4)
Where TTx , TRx and TILS correspond to overall times given in seconds for trans-
mission, reception and Idle-Light sleep respectively and obtained from the assumed
transmission times according to [11] and the packet flow shown in Fig. 3, while Nrep
corresponds to the number of reports per day.
A device of 23-dBm power class with an 5 Wh battery and a base station of
46-dBm power class are used for guard-band and in-band operation modes, while
for stand-alone operation mode the power class of the base station is only 43 dBm
according to [11]. Based on the assumed transmission times given in [11] and using
the formulas (3) and (4), the evaluated battery lifes to achieve MCL of 164 dB in
in-band, guard-band and stand-alone operation modes are 11.4, 11.6 and 11.8 years
respectively. Knowing that the 5G mMTC requires battery life beyond 10 years at
the MCL of 164 dB, supposing an energy storage capacity of 5Wh [8]. Therefore,
NB-IoT achieves the targeted battery life in all operation modes.
Furthermore, to increase battery life further, the Narrowband Wake-Up Signal
(NWUS) that was introduced in 3GPP Rel-15 can be implemented, since it allows
the UE to remain in idle mode until informed to decode NPDCCH channel for a
paging occasion, thereby achieving energy saving.
The 5G mMTC target on connection density that is also part of the International
Mobile Telecommunication targets for 2020 and beyond (IMT-2020), requires the
support of one million devices per square kilometer in four different urban macro
scenarios [8]. These scenarios are based on two channel models (UMA A) and (UMA
268 A. Abou El Hassan et al.
B) and two distances of 500 and 1732 m between adjacent cell sites denoted by ISD
(inter-site distance) [12].
Based on the simulation assumptions given in Table 5 and the non-full buffer
system level simulation to evaluate connection density according to [13], Fig. 5 shows
the latency required at 99% reliability to deliver 32 bytes payload as a function of the
connection attempts intensity (CAI) to be supported, corresponding to the number
of connection attempts per second, cell and PRB. It should be noted that the latency
shown in Fig. 5 includes the idle mode time to synchronize to the cell and read the
MIB-NB and SIB1-NB by the device.
Knowing that each device must submit a connection attempt to the system peri-
odically, we can calculate the connection density to be supported (CDS) per cell area
using the following formula:
C AI · C A P
C DS = (5)
A
Where CAP is the periodicity of device’s connection attempts
√ given in seconds
and the hexagonal cell area A is calculated by A = I S D 2 · 3/6.
To evaluate the connection density per PRB and square kilometer depicted in
Fig. 6, and corresponding to the overall number of devices that successfully transmit
a payload of 32 bytes accumulated over two hours with the required latency, the CDS
values are obtained from formula (5) using the CAI values of Fig. 5 and a periodicity
of connection attempts of two hours.
As can be seen from Fig. 6, in the two scenarios corresponding to 500 m ISD,
more than 1.2 million devices per PRB and square kilometer can be supported by
an NB-IoT carrier with a maximum 10 s latency. However, only 94000 and 68000
devices per PRB and square kilometer can be supported in the scenarios using the
(UMA B) and (UMA A) channel models respectively with an ISD of 1732 m within
the 10-s latency limit.
Since, in the case of 1732 m ISD, the density of base stations is 12 times lower
than with 500 m ISD. Therefore, this difference in base station density results in
Narrowband IoT Evolution Towards 5G Massive MTC … 269
Fig. 5. Intensity of
connection attempts in
relation to latency
4 Conclusion
To conclude, this paper shows that the five targets of 5G mMTC are fulfilled by NB-
IoT. However, the evaluation results show that the performance achieved is highly
dependent on certain conditions relating to system configuration and deployment,
such as the number of repetitions configured for channels transmission and the density
of base stations, which is significantly dependent on the inter-site distance ISD. Since
the connection density evaluation shows that the target of 5G mMTC is only achieved
if ISD is 500 m. In addition, 3GPP Rel-15 allows 5G NR to coexist with carriers when
270 A. Abou El Hassan et al.
NB-IoT devices are already deployed in massive IoT applications. Thus, NB-IoT can
be considered as a key technology of the 5G system for massive IoT applications,
such as smart metering systems allowing measurement of water, gas and electricity
consumption.
References
Abstract Major advancements and development, in the field of Internet and other
technologies is now prevalent in many modern areas and dominating the way of
life. This in turn has allowed us to connect everyday objects, such as sensing and
actuating devices, to the digital world. These intelligent devices empower the dawn
of state-of-the-art services and applications for a multitude of various domains and
help in improving the complete utilization of the given resources. In this paper, an
interoperable, low cost, energy efficient Internet of Things platform using ontologies
for a smart home is being implemented. The said system promises to controls the
home appliances from anywhere i.e. offers remote operability and also provides
the collected data to Third party programs if need be, so that the Third parties can
creates tailor- made applications and services that cater to the need of the consumer.
The system promises to reduce human intervention, enable automation and provide
access to devices remotely from any place. Also, it allows for the collected data to be
sent to third parties for generation of third- party services, and also for the analysis,
and overall improvement of the system.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 271
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_25
272 L. Varghese et al.
a house into a smart home, for some, could mean buying a wireless speaker, while
it can mean different for others, such as connecting several various products for
creating a so called smart environment or home. Home automation aims to enhance
the comfort, energy consumption efficiency and security in domestic scenarios [6].
We aim to reduce the human intervention, secure access control to home devices
from anywhere, provide smart homes data for application services as well as for
analysis, and improve the utilization of resources [5].
Motivation: As our resources such as electricity, drinking water, food and other
natural resources are getting scarcer and more expensive, there is a need to keep
the utilization of these resources in check. In order to reduce the wastage of these
resources and to ensure the safety of the people within a household, there is a great
scope and need for technologies, such as smart homes, which help to monitor and
save the resources and also try to keep the household secure from many mishaps,
such as fires, theft, etc.
Contribution: This paper is being done in two phases, in the first phase, identification
of the topic, research work on the topic and a simulation is being carried out. In the
second phase the implementation and ontology part of the paper is carried out. The
overall system can further be used in various other applications, with addition of
various other appliances. The system monitors room temperature, detects smoke
(mainly carbon dioxide), it also detects fire and tries to prevent any damages using
other appliances such as water sprinkler, alarm, etc. Further, all these devices can be
remotely controlled through cloud and are secured using an ID and password.
2 Related Work
In [1], the authors, have integrated the semantic characteristics into the smart home
system paradigm and have created an ontology based semantic service model which
they call OSSM and that is what acts as the middleware for their system. In [2],
the authors use a structure based on an ontological model to facilitate the mindless
arrangement of fitting uses, instead of traditionally programming each household.
In [3], the authors have successfully provided an answer for securing households
by taking decisions dynamically with help of intrusive devices. In [4], the authors
suggest a smart home, using Raspberry Pi and IoT, it is done by connecting cameras
and sensing devices into a web app. In [5], the authors aim to reduce the involvement
of humans, ensure secure access to home appliances from any place, also provide
analytical services and help improve the system. In [6], the authors propose an inter-
face through which context awareness and decision making is carried out using
A Fuzzy Ontology Driven Integrated IoT Approach … 273
distance based fuzzy algorithm. In [7], the authors have used a fuzzy logic approach
for real- time checking and analysis of the data collected from temperature sensor,
heart beat sensor and gas sensor. Data collected is used to evaluate if dangers exist.
This IoT based smart home is built using a powerful tool, namely Raspberry Pi 3
Model B+ microcomputer. It is very cooperative when it comes to minimizing the
required system components. This system uses a multitude of sensors such as motion
detector, fire sensor, smoke sensor, etc. All these sensing appliances are interfaced
using the GPIO pins present on Raspberry Pi board. The general architecture of the
overall system can be seen in Fig. 1 below. A simulation of this said architecture has
been carried out in Cisco Packet Tracer tool.
4 Implementation
certain events were triggered, ensuring the minimization of any materialistic loss, as
well as, any loss of life.
We ran the fuzzy ontology for the scenario when there is a gas leak and or a fire
in the house and plotted the graph for the same, only one example is shown in Fig. 2
below, but this was carried out for a dataset of values for temperature and smoke,
and great results were achieved.
The value 27.377, which we got in Fig. 2 above, indicates that our environment is
relatively in a safe zone, but at the same time, caution is to be taken as the environment
is still slightly above the acceptable environment (i.e. 23.5). So a prompt is sent to
the user to ensure if things are alright or not.
The value 21.60 which we got in Fig. 3 below, indicates that our environment is
in a very safe zone, though a little caution is to be expected as you can see, sector 2 is
also slightly shaded. Although the environment is stable, just some minor supervision
is required. This is only the case during the early stages of deployment. Once the
system learns the patterns of the consumer, it will make decisions accordingly.
The value 33.50 which we got in Fig. 4 below, indicates that our environment
is in the cautionary zone as well as in the danger zone. As you can see, sector 2 is
now shaded almost 75% and sector 3 is shaded around 25%. Also, after seeing the
inputs and checking the rules which were created by the user, code can be found
in Appendix A.1, we can infer that the temperature is more than what the user has
deemed cautionary, and it is this value of temperature which has caused the system
to be in sector 2 as well as sector 3. The amount of smoke is within the cautionary
limit. So, both parameters need to be monitored, and until the system is resolved,
preventive measures need to be taken i.e. triggering alarm, opening window/s etc.
Similarly, many other values were read, and he graphs for them were plotted and
inferred from.
Accuracy was used as a metric for the system. Equation 1 below, shows how the
accuracy is calculated.
Correct Outcomes
Accuracy = (1)
Total no. of outcomes
The system was made to run a total of 2046 trials (Total no. of outcomes), out of
which 1957 trials were perfect (Correct Outcomes) while others were wrong, due to
wear and tear of the sensors. These achieved values imply that our system has about
96% accuracy, which is very good.
5 Conclusions
In this paper we have simulated one feasible solution for smart home system using
IoT and fuzzy ontology. One of our primary targets was to find an affordable and
practical solution for a real-time smart home system, and we were able to achieve it
by using the sensing devices along with the Raspberry Pi. Certain events were created
to simulate mishaps which could potentially cause damage to property and life, but
we were successfully able to negate any damage by designing a fast, responsive and
interactive smart home system. The fuzzy ontology part also yielded great results,
which has allowed us to complete the project successfully. This paper can be easily
scaled up or down by addition or deletion of sensors. Furthermore, a more visual
and informative, GUI (Graphical User Interface) can be used, based on the users
requirements, to help understand the system in a better and easy way.
References
1. Yang Y, Wang Z, Yang Y, Li M, Yang Y (2012) An ontology based semantic service model in
smart home environment. In: proceedings of 2010 3rd international conference on computer
A Fuzzy Ontology Driven Integrated IoT Approach … 277
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 279
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_26
280 A. Elmouatamid et al.
1 Introduction
Moreover, the ICTs integration enables various smart and automatic services,
such as smart metering infrastructure, smart control and management for D/R
(demand/response) balance, advanced electricity marketing, and intelligent energy
storage for electrical vehicles integration. However, despite this progress, some
research work stated that the SG is experiencing new challenges. Mainly, the SG
is based on the actual infrastructures of power distribution grids, which are limited
by the unidirectional exchange of electricity [3]. Therefore, face these challenges;
other concepts have been developed together with the revolution of SG, such as the
IoE (internet of energy), the IoT, and the IoS (internet of services) [4]. Especially,
the development and the emergence of smart MG systems could resolve some of the
abovementioned SG challenges. MG could simplify the management of electrical
energy, from centralized to distribute EM. In addition, in MG systems, different types
of energy can be managed locally with the possibility to interconnect different MGs
in a distributed manner.
In this work, a new holistic architecture of smart buildings is presented by
improving the main layers of MG. The proposed MG system presents a smart and
active building that combines ICTs/Big-data infrastructure, RESs/storage systems,
EM/control strategies, and electrical power grids. Mainly, the main component of
the energetic system is presented, which is installed in our deployed MG system. We
focused on the modeling of each component of MG layers, especially for sensing and
measurement using our ICTs/Big-data infrastructure. The main contribution of this
work is to propose open-access tools to build an IoT/Big-Data platform for researches
in MG system. The installed platform collects the data from MG system (e.g., power
flow, temperature, humidity, wind speed) using a set of installed sensors. The gath-
ered data is used to develop predictive control approaches for energy management in
MG system using the new ICTs methods (e.g., machine learning algorithms, IoT).
The MG system is structured into three horizontal layers: passive building layer (e.g.,
building envelope and insulation, architecture design), active building systems layer
(e.g., HVAC system, Lighting), and RESs system layer (e.g., PV, wind, storage).
These layers are monitored by one vertical layer for communication and ICTs inte-
gration. This vertical layer integrates mainly an IoT/Big-Data platform in order to
measure, analyze, predict, and forecast actions depending on the actual and predicted
context. In particular, our MG system is a smart and active building that combines
ICTs/Big-data infrastructure, RESs/storage systems, EM/control strategies, and elec-
trical power grids. This new concept of a building is more interactive for both
consumers and energy producers. In fact, consumers will reduce the cost of their
energy consumption based on the used control approaches, which take into account
the real-time cost of the power and the predictive power generation, for efficient D/R
management [4, 5]. Moreover, this MG structure offers the possibility to integrate
new buildings’ services, such as electrical vehicles, which can be used as a storage
282 A. Elmouatamid et al.
The ICTs have a fundamental role in the performance of MG systems operation and
its practical deployment. In an MG system, a large number of distributed compo-
nents, such as the end-users, generators, and energy storage systems, require a reli-
able, sophisticated, and fast communication infrastructure to reach the mandatory
of such systems. Therefore, effective communication systems should be normalized
to enable real-time exchange of data in MG systems in order to collect the neces-
sary information used for the EM (energy management) of the system. Mainly, the
communication technologies that are used actually in MG systems can be classified
into two categories, wired and wireless technologies. Table 1 presents the different
communication technologies, which can be used in the MG system with the specified
international standards [6, 7].
Therefore, the main advantage of the RS-485 system is the flexibility to transport
the signal. It mains that each sensor is defined by its individual address offering the
possibility to be attached to one cable. Multiple sensors use one port on the data
logger saving significantly the installation cable and data logging costs. Mainly, our
deployed MG system is equipped with component for measuring and collecting both
power consumption and production. In fact, a set of current and voltage sensors are
installed for power measurement as shown in Fig. 2. The system is equipped by an
Arduino, which allows collecting the data from different sensors and send it to a
micro-computer (Raspberry pi) connected directly to the IoT/Big-Data platform. In
fact, sensors transmit analog signals to the microcontroller, which converts them into
numerical data.
The next section introduces our MG system’s architecture. In particular, we high-
lighted the necessity of integrating recent ICTs and IoT/Big-Data technologies for
gathering external and internal data, which have been used to generate predictive
actions (e.g., regulating the room temperature by forecasting the building’s occu-
pancy, ventilation speed variation according to the forecasted CO2 , intelligent and
predictive control of energy flows management using forecasted power production,
consumption, and battery SoC).
The aim was to develop a research testsite integrating the different components of
an MG system. As shown in Fig. 2, the system integrates PV panels, WT, batteries,
and the utility grid connected together in order to supply electricity to the loads in
Towards an IoT/Big-Data Platform for Data Measurements … 285
Approaches developed in the frame of this work are deployed and tested in real-sitting
scenarios in two MG systems. The first, named EEBLab (Energy Efficient Building
Laboratory) is used as a testbed for testing scenarios and validation. The second MG
system is a real house, which is deployed during Solar Decathlon Africa competition,
and it is deployed in Green Energy Park in Benguerir, Morocco. Both MG systems
are used actually as a research platform for MG system development. The main aim is
to measure, collect, analyze, and store the different data in the deployed MG systems.
Mainly, the collected data are, especially, the indoor/outdoor temperature, the indoor
CO2 variation, the weather conditions, and the power production/consumption.
Therefore, the measured data are used to develop different scenarios as the occu-
pant’s detection, the comfort amelioration, and the EM/control for the deployed
RESs/storage systems. In this part, we focus on the different measurements realized
for the energetic system. Regarding the external context, we have deployed a weather
station. In fact, for several scenarios, we need to collect internal and external context
data. We have built a weather station near the WT and PVs in order to have as precisely
as possible the data for wind speed, direction, irradiation, temperature, and humidity.
The weather station was deployed and used to collect the data for real-time visualiza-
tion and processing for further usage by other building’s services and applications.
Moreover, D/R control strategies are therefore required for an efficient management
of energy flows by considering the intermittent RES generation and the delay might
occur between RESs power generation and the actual building’s demand. The main
aim is to develop a control card to test the different studied control strategies for EM.
Unlike existing systems, which are used as a black-box to collect and manage the
energy in MG, the deployed control card allows us to measure, monitor, manage,
and deploy our algorithms [13]. Therefore, to develop an optimal and reliable control
strategy, multiple objective functions should be taken into account, like the continuity
of the power generated to the loads, the electrical energy cost, the smoothing of RESs
production„ and the C/D cycle of the batteries [13]. For that, a control strategy should
be deployed to satisfy the constraints designed by the optimization functions. The
communication infrastructure is employed for total electrical energy measurement
and management. The installed infrastructure provides the autonomous operation
with the required measurements, decisions, and command by gathering data through
the sensors and producing the commands for the Hardwar/Software (Hw/Sw) card
presented in Fig. 4, which is connected to the control switches used in the hybrid
system [10].
For the different scenarios, the proposed IoT/Big-Data platform could be used to
measure different parameters in the MG system. Depending on the studied scenario,
Towards an IoT/Big-Data Platform for Data Measurements … 287
suitable sensors are selected and can be connected to this platform for data collec-
tion, monitoring, and processing. Mainly, the measured parameter is used to control
efficiently the deployed MG system. Moreover, a case study is presented for a large-
scale installation in which the developed control card and the IoT/Big-Data platform
are used to measure and store the data collected by the deployed current and voltage
sensors for the total installed RESs, battery storage system, and load consumption.
As shown in Fig. 5, the green curve presents the power generated from the eight PV
panels for 24 h. This power is calculated by measuring the PV current and voltage
variability during the day, which depends on the weather conditions (e.g., tempera-
ture, irradiance) changeability. At the same time, the battery SoC is calculated using
our battery characterization system installed in the MG system. Moreover, the power
consumption is measured and stored for the same period. These parameters are the
main input for the EM strategy.
However, due to the limited number of pages of this work, the obtained results
are summarized in Table 2 for the different measurements realized in the deployed
MG system focusing on that are used for EM in the deployed hybrid RESs system.
Table 2 presented the obtained results for 10 days of the test. Actually, the house
288 A. Elmouatamid et al.
is used as a research platform and it is planned to deploy the test control scenarios,
tested in a medium-scale, for large and real rural MG systems.
4 Conclusions
Acknowledgments This work is supported by MIGRID project (grant 5-398, 2017–2019), which
is funded by USAID under the PEER program. It is also partially supported by HOLSYS project,
which is funded by IRESEN (2020–2022).
References
1. Kabalci E, Kabalci Y (2019) From smart grid to internet of energy. Academic Press
2. Asaad M, Ahmad F, Alam MS, Sarfraz M (2019). Smart grid and Indian experience: a review.
Resour Pol 101499
Towards an IoT/Big-Data Platform for Data Measurements … 289
Abstract Until now, the Covid-19 pandemic was infected more than 37,167,291
people around the world in 212 countries and territories, killed 1,071,181 people
since its appearance in December in Wuhan, of which more than 7,876,942 people in
the United States, 6,979,434 people in the India and 5,057,958 people in Brazil
(Novel Coronavirus (2019-nCoV) situation reports - World Health Organization
(WHO), 2019 Novel Coronavirus (2019-nCoV) in the U.S -. U.S. Centers for Disease
Control and Prevention (CDC), symptoms of Novel Coronavirus (2019-nCoV) –
CDC, china Travel Advisory - U.S. State Department, accessed January 31, 2020.).
This large number of infected people caused saturation in hospitals and put a big
pressure on the intensive care units and hospital beds. While the medical staff are
fighting to save the lives of these patients, others are playing with these lives by
making cyber-attacks on research centers and health institutions, recently case was
in September 2020, a woman in Germany died during a ransomware attack on the
Duesseldorf University Hospital. Smart hospitals are one of the most affected because
they use its technology (Smart hospitals) that generates a lot of personal and crit-
ical data, these hospitals are equipped with “things” that are used for patient moni-
toring, maintenance, remote operation and control like wearables, smart pills, smart
beds, remote monitoring systems, RTHS (Real-time Health Systems), biosensors,
glucose measurement devices, robots, equipment monitoring devices, and more. In
this paper, we will propose a new secured system model to keep these hospital
systems safe against cyber-attacks by using verification and classification of treat-
ment requests and controlling the patients’ health using “sensors” to ensure the safety
of the requested operations.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 291
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_27
292 Y. Chahid et al.
1 Introduction
The Internet of things continues to enter our lives slowly every day and affects
many areas [2]. Healthcare is one of the richest domains that contain the biggest
connected objects, thermometers, ultrasounds, automatic respirators, glucose moni-
tors, digestible sensors, smartwatch, electrocardiograms and others, have become
connected and letting patients track their health and facilitate the doctor’s work by
automating some tasks like operations, drugs injections or insulin control [3]. Several
hospitals have started to use smart beds, which can be equipped with sensors like
injectors or blood collectors, insulin pens, automatic respirators and others, in order
to facilitate the patient treatment operation and provide great support without the need
for a doctor or nurse intervene [4]. The internet of things in the field of healthcare can
also help patients to take their medicament with the right doses directly from their
homes without having to stay in the hospital, this can also guarantee doctors better
monitoring their patients and having an alert if ever there is the slightest concern
[5]. IoT market in healthcare is predicted to exceed $10 billion by 2024 [6], which
means billions of critical and personal data that will circulate on networks. In 2016,
73 million healthcare devices were connected worldwide. In 2020, they will be 161
million [7]. Growth will mainly be driven by 3 trends: the rising average age of the
world’s population, the prevalence in some countries of diseases requiring regular
monitoring (such as diabetes) and the growing demand for quantifiable fitness solu-
tions. This paper is organized as follows: in the first section, we give a state of the art
of the MIMT attack and the related work already done by other research to improve
security against this attack, in the second section we will present a new model that
can help us to improve security inside of smart hospitals, the last section will discuss
the results obtained.
The large amount of data generated by the IoT in the healthcare field forces us to fear,
because this information attracts attackers and make this data accessible everywhere
in the world if we do not apply strong security against these attacks. According to [8]
Large numbers of domain names related to Covid-19 have been registered in recent
weeks, more than 50% of which are thought to be part of “lure and decoy” operations.
Phishing attempts to have soared by over 600% since the end of February, including
traditional impersonation scams but also business email compromise (BEC) and
extortion attacks [9]. Other e-mail campaigns are asking users to transfer money
to sites claiming to be collecting charitable donations or to fake e-commerce sites
offering personal protective equipment.
Smart Hospitals and Cyber Security Attacks 293
The healthcare domain is not immune to these attacks, below are some events
examples made during the COVID-19:
1. May 2020, the FBI issues Flash Alert MI-000124-MW, covering specific
indicators for COVID-19 phishing email campaigns.
2. April 2020, Attacks/campaigns using COVID-themed lures has continued to
increase over the last week.
3. April 2020, several Android-focused campaigns were observed spreading the
Anubis and Cerberus banking trojan to victims seeking additional information
on Coronavirus in their area.
4. March 2020, a fake email using CORONAVIRUS information disguised as an
alert from the “Department of Health” was propagated.
5. April 2020, fake charity and donation have become more and more frequent
since the onset of the pandemic.
2 Related Work
Many efforts have been performed to develop a secured IoT healthcare system. In
this section, we cite some works that are reviewed and summarized.
Fathi Ibrahim Salih et al. have proposed an IoT security risk management model
for healthcare industry [10], as a solution, this paper proposes an enhanced IoT
risk management model for healthcare with consideration of three risk categories:
Secured Technology, Human Privacy and Trustable Process and Data. Bahar Fara-
hani et al. discuss applicability of IoT in healthcare field by presenting an architecture
of IoT eHealth ecosystem [11]. Behailu Negash et al. [12] presents a chapter focuses
on a smart e-health gateway implementation for use in the Fog computing layer, also,
connecting a network of such gateways, both in home and in hospital use. Simple
healthcare scenarios are presented. The features of the gateway in the Fog implemen-
tation are discussed and evaluated. Nasser S. Abouzakhar et al. [13] have published
a paper that presents different risks and security threats in the IoT, some security
measures and solution models as well as a vulnerability assessment approach for the
IoT in the healthcare field. Florian Kammüller [14] show in his paper, how to derive
formal specifications of secure IoT systems by presenting a process that uses the risk
assessment strategy of attack trees on infrastructure models, his proposed process in
the paper uses two features to refine a system specification until expected security
and privacy properties can be proved. He illustrates, in the final part of his paper, the
stepwise application of the proposed process in the Isabelle Insider framework on the
case study of an IoT healthcare system. Maria Almulhim and Noor Zaman [15] have
published research that proposes a secure group-based lightweight authentication
scheme for IoT based E-health applications, the proposed model provides mutual
294 Y. Chahid et al.
authentication and energy efficient, and computation for healthcare IoT based appli-
cations, which will use elliptic curve cryptography (ECC) principles that provide
mentioned featured of suggested model.
The most dangerous attack in healthcare is that which relies on the use of sensors
(insulin pens, automatic respirators, drugs injection…etc.) to inject higher/lower
doses according to the patient’s health condition, stop automatic respirators or modify
the volume of oxygen transmitted, or even the attack at the same time of remote
surgery, these attacks can cause immediate patient death. To cope with this type
of attacks, we are going to propose a secure system model which is based on the
classification of the requested operation types for patients and to control access or
any treatment according to these categories.
In this table (Table 1), we give an example of attack which is based on the use of
Man-In-The-Middle (MITM) which is very well-known in the attack’s types, MITM
is an attack type when an attacker positions himself in the middle of the conversation
and precisely between a user and an application to imitate one of the parties, making
it appear as if a normal exchange of information is established.
The goal of this attack is to steal personal data: login credentials, credit card numbers
or account details. Victims are generally the financial applications, e-commerce
websites or smart hospitals. The MITM attack come in many types [16].
The proposed model will be positioned between the hospital’s external server and
the client, which in our case is the hospital (Fig. 1).
The system will aim to retrieve requests and be able to process them according to
their severity, the system will play the orchestrator role between any patient’s equip-
ment like electrocardiogram (ECG), Electromyography (EMG), Electrogastrogram
(EGG), Insulin pens…etc.) using tags [17, 18] (Fig. 2).
To explain this algorithm, we will analyze each step (Table 2):
Before continuing reading this article, it should be noted that this paper is a computer
science paper and does not encourage the use of any medication type against COVID-
19 (We used the chloroquine just as an example).
The proposed system is implemented using Python Web APIs and MySQL as
database for storing logs and processed requests. We simulate a smart hospital,
patients, and we regulated the categories as follows:
BEGIN
needed_value 0
isSafe false
levels empty_list
values empty_table_params
levelList Get_category_levels
// The system verify the criticity of the operation
critic verify_criticity(RequestedOperationType, values)
tag_level_action
IF isSafe
do_operation
ELSE
PRINT "Potential critical action”
PRINT "Lanching internal checks”
// Get chloroquine, glucose, tension values of the patient
values Get_values(RequestedOperationType)
IF level 2
needed_value calculate_needed_value
do_action(needed_value)
ELSE
reject_action
send_report_to_doctor
save_operation_logs
END IF
END IF
END
Fig. 3 Patient’s status graphic with injected dose values without system implementation
Table 4 Patient’s status with injected dose values in case of system implementation
Case Requested chloroquine dose Patient’s health
1 1000 Normal
2 5000 Normal
3 5000 Normal
4 120 Normal
5 3000 Normal
6 430 Critic
7 1000 Normal
8 1000 Normal
9 100 Normal
10 2670 Normal
When comparing our results to those of older studies, it must be pointed out that
our proposed model can be integrated with any third system, also, the model can
improve the security of smart hospitals against MITM attacks because he is the last
who check requests and it is the closest to patient and this can guarantee that no one
can modify data after the check.
Smart Hospitals and Cyber Security Attacks 299
Fig. 4 Patient’s status graphic with injected dose values with system implementation
5 Conclusion
References
1. Novel Coronavirus (2019-nCoV) situation reports - World Health Organization (WHO), 2019
Novel Coronavirus (2019-nCoV) in the U.S -. U.S. Centers for Disease Control and Prevention
(CDC), symptoms of Novel Coronavirus (2019-nCoV) – CDC, china Travel Advisory - U.S.
State Department. Accessed 31 Jan 2020
2. Chahid Y, Benabdellah M, Azizi A (2017) Internet of things security. In: International confer-
ence on wireless technologies, embedded and intelligent systems (WITS). https://doi.org/10.
1109/WITS.2017.7934655
3. Al-Refaie A, Chen T, Judeh M (2018) Optimal operating room scheduling for normal and
unexpected events in a smart hospital. Oper Res Int J 18:579–602. https://doi.org/10.1007/s12
351-016-0244-y
4. Hurysz T, Romisher A, Campbell D, Ottenstein Maia (2020) Utilizing the smart rooms device
to improve patient experience in hospitals. Phase 1. Paper 12. https://jdc.jefferson.edu/si_dh_
2022_phase1/12
5. Gandhi DA, Ghosal M (2018) Second international conference on inventive communication
and computational technologies (ICICCT). https://doi.org/10.1109/ICICCT.2018.8473026
300 Y. Chahid et al.
6. Internet of Things (IoT) in Utility - Global Market Research and Forecast, 2015–2025 Report
ID: HeyReport6861 | Number of pages: 80 | Publish Date: Feb 2019 | Publisher: HeyReport |
Category: ICT and Media
7. Internet of Things (IoT) in Healthcare Market Size, Share & Trends Analysis Report By Compo-
nent (Service, System & Software), By Connectivity Technology (Satellite, Cellular), By End
Use (CRO, Hospital & Clinic), By Application, and Segment Forecasts, 2019–2025, Report
ID: 978-1-68038-857-2
8. https://blog.checkpoint.com/2020/03/19/covid-19-impact-as-retailers-close-their-doors-hac
kers-open-for-business/
9. Phil Muncaster UK / EMEA News Reporter, Infosecurity Magazine https://www.infosecurity-
magazine.com/news/covid19-drive-phishing-emails-667
10. Salih FI et al. (2019) IoT security risk management model for healthcare industry. Malaysian
J Comput Sci [S.l.] 131–144. ISSN 0127-9084. https://doi.org/10.22452/mjcs.sp2019no3.9.
Accessed 10 May 2020
11. Farahani B, Firouzi F, Badaroglu VCM, Constant N, Mankodiya K (2017) Towards fog-driven
IoT eHealth: promises and challenges of IoT in medicine and healthcare. Elsevier. https://doi.
org/10.1016/j.future.2017.04.036
12. Negash B et al. (2018) Leveraging fog computing for healthcare IoT. In: Rahmani A, Liljeberg
P, Preden JS, Jantsch A (eds) Fog computing in the internet of things. Springer, Cham. https://
doi.org/10.1007/978-3-319-57639-8_8
13. Abouzakhar NS, Jones A, Angelopoulou O (2017) Internet of things security: a review of risks
and threats to healthcare sector. In: IEEE international conference on internet of things (iThings)
and IEEE green computing and communications (GreenCom) and IEEE cyber, physical and
social computing (CPSCom) and IEEE smart data (SmartData). https://doi.org/10.1109/iTh
ings-GreenCom-CPSCom-SmartData.2017.62
14. Kammüller F (2019) Combining secure system design with risk assessment for IoT health-
care systems. In: IEEE international conference on pervasive computing and communications
workshops (PerCom workshops). https://doi.org/10.1109/PERCOMW.2019.8730776
15. Almulhim M, Zaman N (2018) International conference on advanced communication tech-
nology (ICACT). https://doi.org/10.23919/ICACT.2018.8323802
16. Chahid Y, Benabdellah M, Azizi A (2017) Internet of things protocols comparison, architecture,
vulnerabilities and security: State of the art. In: ACM international conference proceeding
series, a65
17. Kannouf N, Labbi M, Benabdellah M, Azizi A (2018) Security of information exchange
between readers and tags. In: Maleh Y, Ezzati A, Belaissaoui M (eds.) Security and privacy in
smart sensor networks, pp 368–396. IGI Global, Hershey. https://doi.org/10.4018/978-1-5225-
5736-4.ch016
18. Kannouf N, Douzi Y, Benabdellah M, Azizi A (2015) Security on RFID technology. In: 2015
international conference on cloud technologies and applications (CloudTech), Marrakech, pp
1–5. https://doi.org/10.1109/CloudTech.2015.7336997
19. Eddy M, Perlroth N (2020) Cyber Attack Suspected in German Woman’s Death, New York
Times, 18 September 2020
Design of an Automatic Control and IoT
Monitoring System for Dual Axis Solar
Tracker
Fatima Cheddadi, Hafsa Cheddadi, Youssef Cheddadi, Fatima Errahimi,
and Ikram Saber
1 Introduction
Due to the high demand of energy and the ever-increasing use of fossil resources,
the whole world has led to the development of the alternative solution and have
moved towards using renewable energy such as solar, wind, biomasses and hydraulic
energy to decrease the uses of fossil resources. In addition to meet the growing energy
demand, another advantage of the use of renewable energy is to reduce CO2 emissions
and the earth warming.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 301
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_28
302 F. Cheddadi et al.
In this way, the photovoltaic (PV) energy is one of the cleanest energy resources
and the most ecological type of energy to use [1]. It is directly converted into electrical
energy by the use of PV modules. However, the harvested power mainly depends on
climatic variations namely the temperature and irradiance, where the need to track
the maximum power point (MPP). For this purpose, there are two strategies to extract
the maximum power. The first one is an electronic strategy based on the called MPPT
methods that try to find optimal operation points of the PV generators without any
movement. The second is based on electro-mechanical trackers that track the sun
movement during the day.
In the literature, several researchers have developed and implemented different
types of maximum power point tracking controller. These controllers differ from
each other in terms of many parameters, namely the number of required sensors,
convergence rate, rapidity of tracking the MPP when the climatic condition suddenly
change, complexity, and the cost of implementation [2].
Some of the MPPT strategies are based on adjusting the duty cycle of the DC-DC
converter [2–8]; the duty cycle plays the role of an adaptation stage between the load
resistance seen by the PV system and the internal PV module.
The other strategies are based on the implementation of solar trackers [9, 10],
where there are two main families of solar trackers: the passive ones and the actives
that include the single-axis and double-axis trackers.
The energy supplied by the PV panel is highly dependent on the amount of sunlight
absorbed by the panel. This amount depends on the orientation of the panel in relation
to the sun. In order to collect the maximum energy, the PV panel should be constantly
oriented perpendicular to the sunlight [11].
Every day this angle decreases or increases, the surface area (m2 ) of the panel
exposed to the rays decreases and therefore the power produced decreases, hence
the importance of the orientation of the panels in relation to the position of the sun
[12]. Another factor that influences the performance of the PV panel is the tilt angle,
which is the angle formed by the plane of the solar panel in relation to the horizontal.
The purpose of PV monitoring systems is to offer continuously a clear information
to improve energy efficiency, many researchers have proposed a monitoring system
of PV station, which help also on fault detection [13–15]. Authors in [16] have
presented a real-time monitoring system, which contain an alert system to a remote
user, when there is a deviation of solar power.
The main use of the system of monitoring in this work is to highlight the effec-
tiveness of the dual axis solar tracker by comparing the measured power from the
system of tracking with a fixed system.
The paper is organized as follows: Sect. 2 presents the proposed tracker. Section 3
presents the PV monitoring system. Results and a discussion are described in Sect. 4.
The paper is concluded in Sect. 5.
Design of an Automatic Control and IoT Monitoring System … 303
In order to collect the maximum amount of energy, the PV panel must be constantly
oriented perpendicularly to the sun’s rays; this can be ensured by a solar tracking
system allowing to follow the sun throughout the day.
This method is based on light sensors or photo-resistors to determine the position
of the sun at any time, allowing instantaneous tracking, and including optimal panel
orientation. In this type of techniques, tracking is carried out independently of the
zone in which the system will be located.
The two types of solar trackers that exist are either single-axis or double-axis,
which are either are based on the same principle of pursuit illustrated in Fig. 1.
The technique used for this study is based on the difference in incident illumination
on photo-resistance or LDR (Light-Dependent-Resistor) light sensors separated by
opaque walls and placed on the PV panel. The control and command circuit compares
the signals emitted by the sensors by calculating their differences and then sends
pulses back to the motor to reposition the panel perpendicular to the sunlight.
LDR1
M1
Vertical
LDR2 System of
control
LDR3 (ESP32)
M2
Horizontal
LDR4
LDR LDR
Top left top right
LDR LDR
bottom bottom right
left
Start
Yes
| DiffAzim | Stop the left-right servomotor
NO
Yes
The left-right servomotor moves
DiffAzim > 0
the panel to the right.
NO
NO
Yes
| DiffElev | <50 Stop the up-down servomotor
NO
Yes Up-down servomotor moves
DiffElev > 0
the panel upwards
Starting from the previous general principle, to ensure a complete tracking of the
solar path, 4 LDRs distributed in the 4 corners are used.
The technique used consists in checking the equality of illumination of four light
sensors distributed in the four quadrants represented in Fig. 2.
The operating principle is based on the tracking algorithm described in the
following flowchart Fig. 3.
In order to prototype our solar tracker Fig. 4, we use the 3D printing, which
is recognized as one of the biggest technological revolutions in the world today.
It represents a very important new technique for the manufacturing processes of
three-dimensional solid objects.
In this part, we will present the components necessary for the realization of the proto-
type. Some components collect data (INA219 as a current/voltage sensor, BH1750
as an illumination sensor, DHT11 as a temperature sensor). Other components act as
microcontrollers that process and analyze the data and provide wireless connectivity
(dual core ESP32), and for monitoring PV production, a web server is embedded to
store the daily collected data using the sensors to follow the metrics of our monitoring
system (ThingSpeak).
In this proposed conceptual system, the ESP32 DEVKITV1 board acts as a micro-
controller that processes incoming data from various sensors. The ESP32 is a series
of low-cost, low-power systems with Wi-Fi and Bluetooth integrated in a single chip.
This card is chosen because it reduces the cost of the monitoring system. The
ESP32 card is based on the microcontroller Xtensa LX6 dual core 32-bit processor
Tensilica, it has built-in Wi-Fi and Bluetooth, it runs 32-bit programs. The clock
frequency can go up to 240 MHz and has a RAM of 512 KB. It also offers a wide
variety of peripherals, such as: CAN, CNA, UART, SPI, I2C, it is equipped with an
306 F. Cheddadi et al.
Current
100
Current (mA)
80
60
40
20
0
13:26 13:33 13:40 13:48 13:55 14:02
Time
Voltage
14
12
Voltage (V)
10
8
6
4
2
0
13:26 13:33 13:40 13:48 13:55 14:02
Time
Power
1200
1000
Power (mW)
800
600
400
200
0
-20013:26 13:33 13:40 13:48 13:55 14:02
Time
integrated Hall effect sensor and an integrated temperature sensor.so, it can support
monitoring and tracking programs.
The developed monitoring system is capable to measure in real time the energy
produced by a photovoltaic station composed of 3 polycrystalline PV panels placed
Design of an Automatic Control and IoT Monitoring System … 307
in parallel capable to generate 1.6 W of power and 12 V of voltage for each one at
the STC. To follow up and verify their correct operation and to analyze and have an
overview of the entire system, the figures above present the results of the monitoring
on 28/10/2020 in Fez Morocco.
Three parameters namely Power, current, and voltage, are monitored using the
experimental prototype Fig. 5. This latter is connected using WIFI and visualized
via ThingSpeak platform.
Between 13:50 and 13h55 the panels was exposed to the shadow which results
in a decrease in the amount of irradiation measured and consequently a decrease in
voltage, current and power have been detected.
The main objective of this work is to highlight the effectiveness of the dual axis solar
tracker by comparing the measured power from the tracking system with the fixed
system.
We have imitated the behavior and the movement of the sun, which is emulated
by a lamp in five positions as shown in the Fig. 6, for both systems (fixed and with
tracking system) at the temperature 21 °C.
We have taken measurements in the darkness. Figure 7 shows the electrical
measurements in real time, namely the current, voltage and power.
Figure 7 displays the dashboard with different widgets to present and visualize
the results of data collected using the solar tracker (blue curve) and compare them
with fixed system (red curve).
As it is mentioned, the system of control integrates an active method of tracking
to follow and track in real time the sun position in order to maximize the generated
power from the panel.
As it is shown in Fig. 7 and Table 1, even if the position of the sun changes, the
solar tracker is able to follow its position while extracting and keeping the maximum
power. While in case of using the fixed station, the power is maximum only when
the sun is in position 3 which corresponds to the midday productivity.
The increase of energy extracted by the solar tracker could be calculated by the
Eq. 1:
25
20
15
Power (mW)
10
0
20:02 20:09 20:16 20:24 20:31 20:38 20:45 20:52 21:00 21:07 21:14
-5
Time
4
Voltge (V)
0
20:02 20:09 20:16 20:24 20:31 20:38 20:45 20:52 21:00 21:07 21:14
Time
5
Current (mW)
0
20:02 20:09 20:16 20:24 20:31 20:38 20:45 20:52 21:00 21:07 21:14
Time
Fig. 7 Comparison between current, voltage and power monitored using solar tracker and fixed
system
Design of an Automatic Control and IoT Monitoring System … 309
Table 1 Power monitored from fixed system and using the proposed solar tracker
Position POS 1 POS 2 POS 3 POS 4 POS 5 Total
Fixed system (P1) 0 mW 4 mW 22 mW 4 mW 0 mW 30 mW
Solar tracker (P2) 20 mW 22 mW 22 mW 21 mW 18 mW 103 mW
P1 − P2
e f f iciency = × 100 = 70.87% (1)
P1
We found the sum of power extracted by the fixed-axis by the three panels is equal
to 30 mW, and the sum of power generated using the proposed dual axis sun tracker
is equal to 103 mW, which means that the average efficiency of the used tracker is
equal to 70.87%.
5 Conclusion
The objective of this work was to present the results of the implementation hard-
ware of the active tracking method, based on the difference in illuminance of the
light collectors. The proposed sun tracker ensure a continuous maximum of power
extraction.
We implemented in parallel a real-time monitoring system to supervise remotely
the performance of the prototyped sun tracker and to validate the use of solar tracker.
This latter shows good results in comparison to the fixed-axis PV panels with an
important increase in average of 70.87%.
References
6. Bayod-Rújula Á-A, Cebollero-Abián J-A (2014) A novel MPPT method for PV systems with
irradiance measurement. Solar Energy 109:95–104. https://doi.org/10.1016/j.solener.2014.
08.017
7. Bukar AL, Tan CW (2019) A review on stand-alone photovoltaic-wind energy system with fuel
cell: system optimization and energy management strategy. J Clean Prod. 221:73-88. https://
doi.org/10.1016/j.jclepro.2019.02.228
8. Bradai R et al. (2017) Experimental assessment of new fast MPPT algorithm for PV systems
under non-uniform irradiance conditions. Appl Energy 199:416-429. https://doi.org/10.1016/
j.apenergy.2017.05.045
9. Bounechba H, Bouzid A, Snani H, Lashab A (2016) Real time simulation of MPPT algorithms
for PV energy system. Int J Electr Power Energy Syst 83:67–78. https://doi.org/10.1016/j.ije
pes.2016.03.041
10. Awasthi A, et al. (2020) Review on sun tracking technology in solar PV system. Energy Rep
6:392-405. https://doi.org/10.1016/j.egyr.2020.02.004
11. Hafez AZ, Yousef AM, Harag NM (2018) Solar tracking systems: technologies and trackers
drive types – a review. Renew Sustain Energy Rev 91:754–782. https://doi.org/10.1016/j.rser.
2018.03.094
12. Mousazadeh H, Keyhani A, Javadi A, Mobli H, Abrinia K, Sharifi A (2009) A review of prin-
ciple and sun-tracking methods for maximizing solar systems output. Renew Sustain Energy
Rev 13(8):1800–1818. https://doi.org/10.1016/j.rser.2009.01.022
13. Adhya S, Saha D, Das A, Jana J, Saha H (2016) An IoT based smart solar photovoltaic remote
monitoring and control unit, p 5
14. Lopez-Vargas A, Fuentes M, Vivar M (2019) IoT application for real-time monitoring of solar
home systems based on ArduinoTM with 3G connectivity. IEEE Sensors J. 19(2):679–691.
https://doi.org/10.1109/JSEN.2018.2876635
15. Madeti SR, Singh SN (2017) Monitoring system for photovoltaic plants: a review. Renew
Sustain Energy Rev 67:1180–1207. https://doi.org/10.1016/j.rser.2016.09.088
16. Cheddadi Y, Cheddadi H, Cheddadi F, Errahimi F, Es-sbai N (2020) Design and implementation
of an intelligent low-cost IoT solution for energy monitoring of photovoltaic stations. SN Appl
Sci 2(7):1165. https://doi.org/10.1007/s42452-020-2997-4
Analysis Jamming Attack Against
the Protocol S-MAC in IoT Networks
Imane Kerrakchou , Sara Chadli , Mohamed Emharraf ,
and Mohammed Saber
Abstract The Internet of Things (IoT) is one of the growing developments of the
recent period that has to a great extent pulled in industry and the scholarly community.
Life sans IoT is indispensable. In order to disperse any questions about its far-reaching
reception, IoT positively requires both actually and consistently right answers to
guarantee fundamental security and protection. This paper gives a brief description
of IoT and the different technologies utilized, for example, WSN, and inspects known
attacks against remote sensor systems like a Jamming attack of DoS type. Among the
current MAC protocols, the S-MAC protocol is proposed to analyze the performances
of the WSN in case of a Jamming attack using the OMNeT++ test system. Distinctive
application scenarios have been assessed. Performance parameters, for example,
consumed energy, lifetime, system loading, and packet reception rate are the principal
factors taken in our investigation.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 311
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_29
312 I. Kerrakchou et al.
1 Introduction
2 IoT Security
The general objective is to protect the whole system that represents an IoT installation
against a wide range of attacks. The most granular security prerequisites frequently
alluded to as security characteristics [6], are:
– Confidentiality: A concept to ensure that information can only be read by
authorized persons.
– Availability: the objective is to guarantee the continuity of network services
against DoS attacks.
– Integrity: Set of means and techniques to limit the modification of data to
authorized persons.
– Privacy: the objective is to prevent the disclosure of confidential information to
malicious entities and the illegal collection of sensitive information about entities.
In the course of the most recent times, the IoT network was confronted with various
attacks that made the makers and consumers aware of the need to utilizing IoT
equipment with more caution. This section explains the various types of attacks in IoT,
which are physical and cyber-attacks where this last comprise of active and passive
attacks (see Fig. 1). Cyber-attack is a danger that aims at diverse IoT equipment in a
wireless system by breaking the network for manipulation of the user’s information
(i.e., delete, destroy, steal, corrupt). Furthermore, physical attacks allude to the attacks
that interrupt the service and physically harm the IoT equipment.
The following paragraph presents the two kinds of cyber-attacks according to the
gravity in IoT system:
– Active attack: in order to interrupt certain services and manipulate the configura-
tion of the system, an active attack happen by an intruder that gets to the network
and its relating information. There are various methods of attacking the security
of IoT devices, including disruptions, modifications, and interventions. Among
active attacks, we can cite Jamming, Sybil, DoS, MITM, Spoofing, etc.
– Passive attack: this attack attempts to accumulate user information with no his
assent, using this data to break the security of his private information. Traffic
analysis and Eavesdropping are the two fundamental elements to play out a passive
attack using an IoT system.
This subsection presents the generic IoT architecture, which consists primarily of
three layers (see Fig. 1), in order to describe the possible attacks according to each
layer. The IoT surface attacks are classified as follows [7]:
– Perception surface attacks: the attackers may easily penetrate the physical layer
of the IoT system, because the Micro-controllers, Sensors, actuators are a few
pieces of the physical system that are utilized for communication, identification,
and collecting information, and these elements are very susceptible to Jamming,
radio interference, and DoS. Furthermore, physical attacks are the most disturbing
for perception surfaces [8].
– Network surface attacks: In IoT systems, the physical devices are associated
through network services, including wireless networks. This wireless sensor
network is a primary objective for various forms of attacks because user infor-
mation moves overtly with no solid secure protocol. Attacks on the surface of
networks are likely to occur through MITM, DoS, Sybil, Routing, Spoofing, and
so on.
– Application surface attacks: The application layer manages various types of
applications, which allow remote control, and access to IoT devices. This IoT
equipment is linked to the network with clouds and servers. They utilize software-
based applications that allow malicious developers to dispatch distinctive malware
attacks in order to get to IoT devices with or without user authorization. Therefore,
attacks targeting this layer are DoS, Data corruption, etc.
For wireless sensor networks, the MAC protocols in which nodes have a sequence
of active and inactive periods are called activity rate protocols. The activity rate
represents the percentage of time spent in the period of activity. Protocols with activity
rate may be classified into two main types. The first type is that of synchronous MAC
protocols, in which all nodes have their period of activity simultaneously or planned.
The second category is that of asynchronous MAC protocols, in which the nodes
have activity periods independent. There are many MAC protocols, among which
Analysis Jamming Attack Against the Protocol S-MAC in IoT Networks 315
we can cite: S-MAC, T-MAC, B-MAC, etc. In this document, we will talk about
the S-MAC, which is the synchronous protocol. Nodes switch between periods of
activity and periods of inactivity by a fixed action rate. During the active period, a
node can send and get data. The inactive period corresponding to the deactivation of
the radio transmitter/recipient to spare energy. The active period is itself divided into
a synchronization period utilizing a SYNC frame, and a data trade period. The nodes
keep up a table with the wake-up times of their neighbors. At the point when a node is
active, it listens to the channel for a while. If it hears a SYNC frame, it synchronizes
to the schedule of the node transmitting this SYNC frame. If it hears no frame after
this time, it embraces its schedule and communicates it to its neighborhood. All
nodes that follow this schedule form a virtual cluster. At the point when a node has a
frame to transmit, it utilizes a reservation mechanism such as RTS/CTS (for Request
To Send/ Clear To Send) and ACK (Acknowledgement) based on the IEEE 802.11
norm (see Fig. 2) [9].
Jamming is a type of attack that was planned to intentionally intervene with the typical
functioning of the wireless device, at the perception layer, all the more exactly in the
MAC layer, by saturating the canal through the injection and persistent transmission
of data or control package, causing abnormalities in the transmission and reception
of data. What’s more, this attack reduces the performance of the system by increasing
power consumption, memory, bandwidth, etc. Generally speaking, this sort of attack
is associated with Denial of Service attacks (DoS) [10].
However, the Jamming attack could be classified according to the level of knowl-
edge of the protocol used to apply it. Without knowledge of the protocol utilized in
the network, attacks are very restricted in their effects on network performances. On
the other hand, knowing the communication protocol utilized in the system can be
destructive to a WSN. With complete information of the MAC protocol, an attacker
can create big traffic to increase the most extreme effect from Jamming attacks.
316 I. Kerrakchou et al.
In our situation, we are going to launch two types of attacks. First, we assume that the
attacker does not know the protocol utilized. For this reason, he will use DATA packets
to launch this attack. Therefore, he starts replaying these data packets throughout the
network aiming at the sink. Secondly, we suppose that the attacker knows the protocol
used in the network. He analyzes the traffic to determine the protocol utilized in the
system, which is the S-MAC. Then, he starts replaying explicit data messages by
targeting the sink. In this case, the attacker utilizes retransmission of DATA packets
correctly programmed in the S-MAC protocol in order to waste more energy from the
network. Moreover, even in the case of Jamming, as indicated by the MAC protocol,
the lifetime of a WSN can be long due to the idle cycle that disables the nodes,
forcing the attacker to scrambling the system for all simulation time.
First of all, the simulator used in our work is OMNeT++. It is a powerful network
simulator and especially useful when designing and validating new protocols, and
exploring new or exotic scenarios.
In this simulation experience, two scenarios are recreated that have almost the
same qualities with the difference that the first scenario (see Fig. 3) is a classic
situation with no malicious nodes or attacks on the network. We talk about the
normal case.
The second scenario (see Fig. 4) represents the attack of the jammer on the nodes.
This scenario contains one jammer that injects unauthorized traffic into the network
and affects the WSN, which has no particular mechanism for detecting or fore-
stalling Jamming attacks. In this scenario, we simulate the two types of attack that
we explained in the paragraph above. In the first type (no knowledge of the protocol),
the attacker continuously transmits DATA packets in the transmission channel. In the
second type (knowledge of the protocol), the attacker synchronizes with all the nodes
and continuously transmits DATA packets during the activity period.
The jammer used in this scenario is a no-mobile jammer that allows DATA
messages to be transmitted to the sink in order to inject a specific number of packets
into the network and subsequently cause it to fail. These Jamming attacks are one
of the most effective types of Jamming because they reduce the life of the network.
The jammer parameters are shown in Table 1 below:
The simulation model examined utilizes a star topology, where communication
takes place between the nodes and the sink. The three simulations consist of 11 fixed
nodes. Every node is equipped with two AA batteries, enough for a long working
span. The duration of the simulation is set at 100 s. So as to identify the effect of the
attack, the simulation time stays consistent as well as the total number of nodes for
both scenarios. The simulation parameters are shown in Table 2 below:
The principal objective for simulating scenario 1 is to determine the current state of
the network under ordinary circumstances, which will allow us to make comparisons
with scenario 2 and distinguish the effect of a Jamming attack on the network.
However, the goal of the two attack simulations is to determine the degree of the
Jamming attack effect on the network performances.
Table 1 Simulation
Parameters Values
parameters for Jammer node
No. of jammers 1
Transmit Power (mW) 57.42
Trajectory Fixed
Data rate (pps) 50
Data packet size (bytes) 250
End time (s) End of simulation
318 I. Kerrakchou et al.
Table 2 Simulation
Parameters Values
parameters
Simulation time (s) 100
Simulation area (m) 40 × 40
Topology Star
No. of nodes 11
Mobility model No mobility
Transmit power (mW) 46.2
Data rate (pps) 5
Data packet size (bytes) 100
End Time End of simulation
Protocol S-MAC
SYNC packets size (bytes) 11
Frame time (ms) 610
Contention period (ms) 10
In this section, network performance behavior is analyzed with and without Jamming
attacks. For the second scenario, both attack simulations used a higher packet size,
packet rate, and transmission power compared to scenario 1 where no attack is imple-
mented. Based on the results of the simulation experiences, the following outcomes
are obtained.
Figure 5 presents the consumed energy by all the nodes, for the three situations.
For the normal case, we notice that the S-MAC has better performance. The energy
consumption is normal, which means that the network is functioning properly. At the
point when the network is subject to a Jamming attack, with or without knowledge of
the protocol used in the network, the system has a much higher energy consumption
compared to the result obtained in scenario 1.
In our example, the role of the sink (node 0) is to receive and collect packets from
the entire network, so it does not transmit data packets. What explains its high energy
consumption in the case of the no knowledge of the protocol compared to the normal
case, is the reception of the traffic generated by the attacker. On the other hand,
we notice that in the case of the knowledge of the protocol used, the consumption
is much higher because of the reception and processing of a large number of fake
packets sent. The knowledge of the protocol used in the network allows the attacker
to behave like an ordinary node, which allows him to deceive the sink and let it
process its fake packets.
Moreover, the continuous sending of these data packets allows an attacker to
force other nodes to send more frames than expected since a communication failure
is usually followed by several more attempts. With each new attempt, the nodes
consume even more power.
Analysis Jamming Attack Against the Protocol S-MAC in IoT Networks 319
Once the power consumption is high, network life will decrease. The life of a
network is defined as the average time between the deployment of the network and
the moment when the power supply to the nodes is exhausted. Figure 6 shows the
estimated lifetime of the network for the three situations. In the case of the attack,
the life of the network is much shorter than in the ideal case. This is due to the
Jamming attack, which increased traffic, resulting in wasted energy and therefore
reduced network life.
Figure 7 represents the number of DATA packets sent by all the network sensors
and received by the Sink. We can see that sending and receiving packets work
correctly in the normal case. Once the attack is implemented, the number of packets
sent and received decreases.
This decrease in the number of packets sent is due to the high throughput broadcast
in the network by the malicious node. The attacker increased traffic and made the
transmission channel busy, which prevented other nodes in the network from sending
their packets properly and communicating with the Sink. As far as packet reception
is concerned, we can see that the Sink can no longer correctly receive the packets sent
to it by the legitimate nodes due to its occupation with receiving the fake packet and,
above all, with processing the fake packets in the case of knowledge of the protocol
used. Which explains the gross decrease in the reception rate. Figure 8 shows the
decrease in packet reception rate by each node.
From the results obtained, it is deduced that both types of attacks harm network
performance. On the other hand, the attack with knowledge of the protocol used
in the system is much more effective compared to an attack without knowledge,
because the attacker, who already knows the protocol used, diffuses the traffic in the
network following all the rules of the MAC protocol used for collision avoidance,
synchronization, etc.
5 Conclusion
Most current IoT security research is concerned with confidentiality and integrity
data, to a great extent disregarding data availability. Without the capacity to secure the
physical medium in which communication happens, sensor networks are susceptible
Analysis Jamming Attack Against the Protocol S-MAC in IoT Networks 321
References
Abstract The Internet of Things (IoT) is one of the latest increasing technologies
that has largely been drawn to the industry and the academic world. Life sans IoT is
completely imperative. In order to dispel any concerns about its far-reaching accep-
tance, IoT needs both real and reliable responses to ensure essential security and
defense. This paper provides a brief overview of IoT and the various technologies
used, such as WSN, and discusses attacks targeting wireless sensor networks, such
as DoS style Jamming attacks. The S-MAC protocol is suggested among the existing
MAC protocols to evaluate the WSN performance in the event of a Jamming attack
utilizing the OMNeT++ test system. Distinctive scenarios have been tested. The main
parameters taken into account in our survey, to analyze network performance, are
the number of packets delivered, energy consumed, network lifetime, and the rate of
lost packets.
1 Introduction
Kevin Ashton proposed the term Internet of Things for the first time in 1999. It didn’t
get much consideration after that for about 10 years. However, massive interest was
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 323
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_30
324 I. Kerrakchou et al.
paid to IoT technologies in 2014 [1]. Any part of our lives could be made simpler
by the Internet of Things (IoT). Its development is motivated by the promise of a
more convenient everyday life, safer road traffic, a more environmentally sustainable
source of energy, and a healthy lifestyle. A broad number of areas of life [2] are
protected by future implementations. The key concept of IoT is to empower structured
communication between objects of normal life (things that have been granted data
identification, transfer, and handling capabilities) utilizing the Internet and to connect
a local object network using uniform protocols to a global network. Intelligent objects
(things) can be interacted utilizing different network technology, like Wireless Local
Area Networks (WLAN), Radio Frequency Identification (RFID), Wireless Sensor
Networks (WSN), and Cellular Services (3G, 4G, LTE, and 5G) [3]. IoT allows
things and people equipped to use any sort of media and service, to communicate
with anybody and anything, everywhere.
There are numerous dangers and vulnerabilities facing the IoT device [4],
including the communication canal, which helps the attacker to manipulate the trans-
mitted data. There are minimal processing power and poor storage space for the
detection device. Due to price specifications, sensor nodes are not furnished with
sealed facilities. If a large number of machine nodes, or a few simple nodes, are
targeted, the life of the network could be limited to a few days. The radio subsystem
is the primary source of energy waste in these sensor systems. For interacting with
the radio, the data link layer, in particular, the MAC (Media Access Control) protocol
is responsible. In this case, the MAC protocol can retain the radio in a non-powered
standby state. Much work in the field of sensor node energy conservation focuses
on MAC protocols. The sensor MAC protocol (S-MAC) will be investigated for this
purpose in this article. In order to compare network performance with and without
the Jamming attack, two cases are proposed.
The remainder of this paper is presented as followed. The security objective of
IoT and the classification of attacks are presented in Sect. 2. A description of the
S-MAC protocol functioning and the attack targeting this protocol are given in Sect. 3.
Through explaining their implementation on the simulator, Sect. 4 discusses the effect
of Jamming attacks against the MAC protocol used in this paper. Finally, a conclusion
is presented in Sect. 5.
2 IoT Security
– Authenticity: For communication between any two nodes, only allowed nodes
can be involved.
– Confidentiality: The leakage of any unauthorized recipient of classified infor-
mation shall be excluded.
– Integrity: When the information is transmitted to IoT devices, data integrity guar-
antees that the originality of the information is not rewritten, copied, or substituted
by the attacker.
– Privacy: In order to preserve privacy, the identity of an individual user should be
covered by a secure IoT device.
– Availability: Approved users can use different services provided by the protocols
of the IoT network to facilitate access to services.
Different IoT systems are separated into different layers: Perception Layer, Network
Layer, and Application layer [6]. Every layer has a feature of its own. To form a full
IoT system, these layers work together. In this section (see Fig. 1), we briefly discuss
the functioning of the IoT layers with the most common security attacks targeting
each of them [7].
– Perception layer: the sensors are used in this layer to detect and acquire infor-
mation about the environment in which they are implemented. The protocols
related to communication between IoT network devices are included in this layer.
In terms of battery energy, storage resources, and computing capability, IoT
devices are resource-restricted. Among the attacks targeting this layer, we can
find Eavesdropping, Jamming attack, physical attacks, etc.
Fig. 1 Layered
classification of IoT attacks
326 I. Kerrakchou et al.
S-MAC is a protocol based on contention, where time is fragmented into large frames.
Two phases are composed of each frame (see Fig. 2):
– Active period: where the node can connect with neighboring nodes, uses control
packets like SYNC (synchronize), RTS (Request To Send), CTS (Clear To Send)
and, ACK (Acknowledgment) and sends DATA messages.
– Sleeping period: of which the node radio is switched off to prevent loss of energy.
S-MAC applies what is referred to as virtual clustering in order to achieve some
coordination between nodes, in which these last regularly transmit special SYNC
packets at the beginning of every frame to ensure that the node and its neighbors
wake up simultaneously. Virtual clusters are created by neighboring nodes and a
shared sleep schedule is set up [8].
The architecture of the system we have proposed consists of a base station (SINK),
which is used to collect and process data centrally, and a set of nodes that allows
packets to be sent. The system will present a Jamming type attack. In our case, it
is assumed that the attacker knows the protocol used in the network, which is the
SMAC. Once the malicious node is implemented in the network, the attacker analyzes
the traffic to determine the protocol used. Then, he will continuously transmit a large
number of RTS (Request To Send) control packets in the transmission canal. Here,
two situations will be studied. In the first situation, the attacker will target an ordinary
node other than the SINK. As soon as the malicious node is installed in the network, it
will look for the node closest to it and which belongs to this system. Once detected, the
Jammer node will start attacking this closest node by sending a lot of control packets
generating high traffic. In the second situation, the malicious node will attack only
the SINK. As a result, the flow of packets sent by the attacker will be transmitted to
the base station.
4.2 Simulation
The simulator used in this document is the OMNeT++ simulator. We use a set of 19
fixed nodes, where node 0 is the Sink. The simulation period is set to 200 s. We use
a star architecture in which all the nodes send their packets to the base station. The
initial energy of each node is 18720 J. The protocol used is SMAC. The transmission
power is fixed at 36.3 mW and the number of packets sent equals five packets per
second. The table below represents the simulation parameters (Table 1).
328 I. Kerrakchou et al.
Table 1 Simulation
Parameters Values
parameters.
Simulation time (s) 200
Simulation area (m) 60 × 60
Topology Star
No. of nodes 19
Mobility model No mobility
Transmit power (mW) 36.3
Packet rate (pps) 5
Data packet size (bytes) 100
End time End of simulation
Protocol S-MAC
SYNC packets size (bytes) 11
RTS packets size (bytes) 13
Frame time (ms) 610
Contention period (ms) 10
In our simulation, three scenarios are proposed. The first scenario (see Fig. 3)
represents a network without attack, i.e. no malicious node is implemented in the
system. The nodes synchronize with each other and then transfer data packets to the
base station.
Simulation and Analysis of Jamming Attack in IoT Networks 329
Fig. 5 Scenario
3—Network with Jamming
attack_SINK
Table 2 Simulation
Parameters Values
parameters for Jammer node
No. of jammers 1
Transmit power (mW) 57.42
Trajectory Fixed
Packet rate (pps) 30
Protocol SMAC
SYNC packets size (bytes) 11
RTS packets size (bytes) 200
Frame time (ms) 610
Contention period (ms) 10
End time (s) End of simulation
The second scenario (see Fig. 4) represents the Jamming attack by implementing
a malicious node, which is node 19, in the network. In this scenario, we will simulate
the attack that targets an ordinary node, which is node 18, of the network as explained
in Sect. 4.1.
In our experiment, we added a third scenario (see Fig. 5) which also represents
the Jamming attack in the network with the difference that in this third scenario, the
attacker targets the base station.
The second and third scenarios have the same values of the parameters of the
attacker node (jammer node), in order to study the difference in the impact of each
of these attacks.
The Jamming used in our simulation is represented by a fixed node, which allows
the sending of a large number of control packets (RTS) with a sending rate equals to
30 packets per second. The transmission power of 57.42 mW and a control packet
size (RTS) equals to 200 bytes. The table below shows the parameters of the jammer
node (Table 2):
330 I. Kerrakchou et al.
This section describes the performance analysis of the three proposed scenarios. The
objective of the simulation of scenario 1 (normal case) is to analyze the behavior
of the network in the absence of undesirable effects, and finally to compare it to a
network under attack. Then, we proposed in scenario 2, the attack of a network node,
and in scenario 3, the attack of the base station in order to compare the severity of
these attacks on the network. The results obtained are as follows:
Figure 6 shows the number of packets sent by each node, in the network, for
the three scenarios. In the normal case (scenario 1) all nodes send almost the same
number of packets. Therefore, the system is working properly. In scenario 2, where
one node in the network (node 18) was attacked, we notice that the whole network
works in the same way as the normal case, except the attacked node that sends only
a very small number of packets. The reason is that the attacker continuously sends
a set of false packets to the attacked node and prevents it from communicating with
the rest of the network. For scenario 3, we notice that the attack has infected the
whole network, which pushed all the nodes to send a very large number of packets.
The reason for this is that the attacker has prevented the sink from responding to the
other nodes, and when the other nodes do not receive any response from the base
station, they keep sending the packets back until they get a response or their batteries
run out.
Figure 7 shows the packet loss rate by each node, on the network, for the three
simulations. For Scenario 1, it can be seen that the packet loss rate is approximately
zero, meaning that all packets that have been sent by the network nodes reach the base
station. Therefore, the network is functioning properly. For Scenario 2, we notice
that the packet loss rate is a bit high compared to Scenario 1 due to the high traffic
generated by the attacker in the transmission canal, which prevented the routing of
some packets. For the attacked node (node 18), almost half of the sent packets were
not routed to the SINK because of the Jamming attack that interrupted the normal
operation of the packet transmission process with huge unnecessary packets. On the
Simulation and Analysis of Jamming Attack in IoT Networks 331
other hand, in scenario 3, almost all the packets sent by all the nodes did not reach
the base station because this attack left the sink occupied by the fake packets sent by
the attacker and made it unable to receive the packets sent by the other nodes.
The graph in Fig. 8 shows the battery energy consumption of each node in Joule.
It can be seen that the energy consumed in scenarios 2 and 3 is higher than that
consumed in the normal case for the entire network because of the attacker, which
generated high traffic in the transmission canal and led the nodes to waste their
energy in unnecessary retransmissions. We notice that the energy consumption of
the attacked node (node 18) for scenario 2, and the base station (node 0) for scenario 3,
is also very high because of the successive sending of large RTS messages generated
by the attacker.
The estimation of the energy consumed in the network, more specifically of the
nodes, also allows estimating the lifetime of the system, which is an essential param-
eter of WSN. Thus, the high energy consumption of each node automatically reduces
the life of its battery and then that of the network as shown in Fig. 9.
332 I. Kerrakchou et al.
5 Conclusion
Most recent IoT security studies are concerned with knowledge on integrity and
confidentiality, generally without regard to the availability of data. Sensor networks
are vulnerable to a progression of possible attacks that concentrate on gradually
depleting the batteries of sensor nodes and making the network unusable in the case
of impossibility to protect the physical canal where contact takes place.
A summary of IoT, specifically WSN, with an emphasis on Jamming attacks
and the power efficiency of wireless systems is given in this article. For evaluating
network efficiency, four parameters were utilized: the number of packets delivered,
energy consumption, the lifetime of the network, and the loss rate of packets. Using
the OMNET++ simulator, the impacts on these performances and the effects of
Jamming attacks on wireless communication networks were analyzed using the S-
MAC protocol.
Future research will analyze a defense framework and find ways to adapt it to
currently existing detection devices in order to build unique methods to defend them
against this form of attack and to prepare new energy-efficient MAC protocols.
Simulation and Analysis of Jamming Attack in IoT Networks 333
References
1. Jha RK, Kour H, Kumar M, Jain S (2020) Layer Based Security in Narrow Band Internet of
Things (NB-IoT). Comput Netw 185:107592
2. Anand S, Sharma A (2020) Assessment of security threats on IoT based applications. Materials
today: proceedings
3. Khattak HA, Shah M, Khan S, Ali I (2019) Perception layer security in the Internet of Things.
Future Gener Comput Syst 100:28
4. Yugha R, Chithra S (2020) A survey on technologies and security protocols: reference for future
generation IoT. J Netw Comput Appl 169:102763
5. Tahsien SM, Karimipour H, Spachos P (2020) Machine learning based solutions for security of
Internet of Things (IoT): a survey. J Netw Comput Appl 161:102630
6. Malik V, Singh S (2020) Evolutionary computing environments: implementing security risks
management and benchmarking. Procedia Comput Sci 167:1171–1180
7. Hassija V, Chamola V, Saxena V, Jain D, Goyal P, Sikdar B (2019) A survey on IoT security:
application areas, security threats, and solution architectures. IEEE Access 7:82721–82743
8. Jagriti, Lobiyal DK (2018) Energy consumption reduction in S-MAC protocol for wireless sensor
network. Procedia Comput Sci 143:757–764
9. Osanaiye O, Alfa AS, Hancke GP (2018) A statistical approach to detect jamming attacks in
wireless sensor networks. Sensors 18:1691
A Brief Survey on Internet of Things
(IoT)
Abstract Today it be a necessity to setting off past the thought that IT as a cost
center and seeing it as a benefit center, on backing of the improvement and method
of the organization. Internet of things can be defined as a propelled standard which
make straightforward gadget a shrewdly gadgets which allow transferring simple
device an intelligent devices capable of exchanging data over a network without
human interaction. It represent a technology that interest research and industry, also
serve on transform the way to think, and work. This paper present an overview on
Internet of things (IoT), its application areas and its challenges in the aim to give a
guideline to future researchers.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 335
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_31
336 F. Z. Fagroud et al.
2 Internet of Things
Internet of things have many definitions, until now there is no standardized definition.
IoT mean system that groups IT devices, mechanic and digital equipments, objects,
animals or people which are interconnected and can communicate data via a network
in a self-sufficient manner, in real time and without any human interaction. It repre-
sent a dynamic network infrastructure which include physical and virtual “things”
which possess unique identifiers, physical attributes, and virtual characteristics into
the information network. Auto-configuration, self-organization, transmit informa-
tion and data, share resources, react and act face to divers situations and changes in
the environment represent the IoT capabilities.
According to the International Telecommunication Union “The Internet of Things
is a global infrastructure for the information society, which provides advanced
services by interconnecting objects (physical or virtual) through interoperable infor-
mation and communication technologies existing or in evolution”. The concept of
“objects” is intended to pose no limit as to the feasible possibilities of implementing
technological tools in everyday elements.
The Cluster of European Internet of Things Research Projects (CERP-IoT) defines
the Internet of Things as: “a dynamic infrastructure of a global network. This global
network has auto-configuration capabilities based on standards and interoperable
A Brief Survey on Internet of Things (IoT) 337
communication protocols. In this network, physical and virtual objects have identi-
ties, physical attributes, virtual personalities, and intelligent interfaces, and they are
seamlessly integrated into the network.” [CLU 2010].
The development of a system where objects are identifiable in a unique
way, communicating with each other and capable of collecting data without any
human intervention is only possible through integrating different technologies.
Various technologies are deployed for the implementation of IoT including: RFID
(Radio Frequency Identification), Sigfox, WSN (Wireless Sensor Networks), Cloud
Computing, NFC (Near Field Communication), and Artificial intelligence.
IoT work in the following way:
• Collect of data: collect of data from the physical world by using sensors
• Communication: transfer of data collected through different types of networks
• Storage and analysis of data: based on many technologies for example Cloud
Computing to storage data collected
• Visualization: Present information to users in a comprehensible way.
IoT is characterized by:
• Heterogeneity: diversity of platforms, hardware and networks used for creating
of internet of things devices. These devices can interact with other devices or
service platforms by various communications technologies.
• Connectivity: allows accessibility and network compatibility. Accessibility is
over a network while compatibility furnish the common ability to consume and
generate data.
• Intelligence: allows objects to respond intelligently to a well-specified situation
and help them to carry out precise tasks
• Dynamic change: about number and state of device
• Interconnectivity: everything will be able to be interconnected with the global
information and communication infrastructure
• Huge scale: the number of devices to manage and communicate between them
will be high compared to devices connected to the Internet
• Detection: the information detected (information of the physical world) allows
the understanding of our complex world
• Security: must be guaranteed to protect the privacy of users.
3 Application Area
IoT is present in different area like health, energy and agriculture which affirm field
of the IoT application is almost unlimited.
Several IoT applications have been developed in different areas since its appear-
ance, for example: Ji-chun Zhao et al. [2] offer a remote greenhouse monitoring
system, which allows automatic control over the ambient temperature, moisture
factors. Tejaswinee A. Shinde and Dr. Jayashree R. Prasad have developed a system
based on IoT, which allows monitoring of animal health. The user can supervise
338 F. Z. Fagroud et al.
the health of the animal from a distance and be alerted by his mobile phone in the
event of any changes in the health status of the animal [3]. Tuan Nguyen Gia et al.
[4] present a system that permit remote control of glucose in real-time and contin-
uously. The presented system help doctors and caregivers to check their patients
anytime, anywhere per a browser or a mobile application. Security is one of the main
concerns, as all health data is considered private personal data. To mitigate the risks
of these applications, strong security mechanisms are needed. Jorge Gomez et al. [5]
have proposed a system which enable students to interact with physical objects in
their environments associated virtually with a learning subject. Khaleel Mershad and
Pilar Wakim [6] have developed a framework for a future LMS (Learning Manage-
ment System: tool for creation, distribution, monitoring and management of various
types of educational and training material) enhanced by IoT capabilities, several
LMS elements will be affected by IoT. Academics, researchers and students are in a
same position to carry on the exploration and advancement of IoT systems, devices,
applications, and services. The deployment of these solutions in our environment
will transform our environment into intelligent environment and conducive to entire
human activity.
A Brief Survey on Internet of Things (IoT) 339
4 IoT Challenges
4.1 Security
Security is an important aspect of IoT and must be guaranteed in the various layers
of the IoT architecture. They are several challenges related to security in internet
of things for example control access, data exchange security, protect the privacy.
Traditional security mechanisms cannot be applied in internet of things which make
necessary the creation of a news paradigms. To solve this issue we can use Machine
Learning, Blockchain, Gateway, Software Defined Networking, symmetric cryptog-
raphy, biometric modalities, wireless device radio fingerprinting to boost the security
and privacy of IoT systems.
Several research are proposed like: FLAUZAC Olivier et al. [7] propose a new
security architecture for IoT (Internet of Things) based on SDN (Software defined
Network), the latter allows to achieve network security in a more efficient and flexible
way. Hui Suo et al. [8] present an analysis of the characteristics and security require-
ments from the IoT architecture based on four layers (perceptual layer, network layer,
support layer and application layer). An Autonomic Agent Model Trust is a model
that serves to reduce security issues, raise reliability and credibility, and afford infor-
mation in IoT environments. It consists of installing Trustworthy Agent Execution
Chip (TAEC), which adduce a profitable software and hardware platform for secured
operation of the agent, on each sensor node, providing a hardware execution envi-
ronment for agents which is secured and autonomous [9]. The work of Markowsky
et al. [10] demonstrate that the search for vulnerable devices in the IoT is possible by
using a different methods: Shodan (find vulnerable Cayman DSL routers), Masscan
(quickly search a large address space for devices vulnerable to the Heartbleed Bug)
and Nmap with PFT (find and connect to vulnerable networked printers).
Most of IoT devices proposed until now are note interoperable which increase data
consolidation costs. There are 4 types of interoperability:
• Technical Interoperability: related to hardware/software components, systems
and platforms
• Syntactic Interoperability: structures and formats of data communicated
• Semantic Interoperability: same content are comprehended by different means
• Organizational Interoperability: organizations may communicate and transfer
data even they use a diversity of information systems on much different
infrastructures, perhaps at different geographic regions and cultures.
To solve the issue of interoperability in IoT many techniques and methodologies
can be used like: Ontology, RDF, SPARQL, IoT HUB. For example: Agarwal et al.
[16] propose a holistic and a light-weight ontology that aims to attain semantic inter-
operability between divers fragmented testbeds (that store data in their proprietary
format) in IoT domain.
IoT devices yield a broad deal of data in real-time and unstructured, which requires
a decentralized structure to store and analyze this huge data. Integration of Cloud
Computing and IoT represent a valid approach that will serve to solve this issue.
Different IoT cloud platforms have been developed to solve various service areas such
as application development, administration of devices and systems, heterogeneity
management, data management, analysis tools and deployment [17]. Cloud4IoT is
a platform for self-acting deployment, arrangement and dynamic configuration of
software components for IoT support and data processing and analysis applications,
enabling plug-and-play integration of novel sensors and dynamic scalability [18].
CloudThings Architecture is a cloud-based IoT platform that supports IaaS, PaaS,
and SaaS to accelerate the application development and administration of the Internet
of Things [19]. This integration faces a set of challenges: Resource assignment,
Protocol Support, Identity Management, Energy effectiveness, Quality of Service
attainment, Data Storage Location, Security and Privacy.
A Brief Survey on Internet of Things (IoT) 341
5 Research Evolution
The opportunities of internet of things are diverse and unlimited which make then
an important research topic that occupies latterly a great importance by researchers.
The amount of research works interested in the IoT thematic increases from year to
year, to affirm this we present an overview on the number of publications included
in two scientific databases (IEEE and dblp). We carried out this study based on two
scenarios:
• First scenario: search for the number of publications including the chosen
keyword
– Keywords used: Internet of Things, IoT, Internet of Things challenge, IoT
challenge, internet of things application, IoT application
– Results:
• Figure 2: show the number of journal papers published in dblp and IEEE
and respond to our request
• Figure 3: show the number of conference and workshop papers published
in dblp and IEEE and respond to our request.
Fig. 3 Number of publication in IEEE and dblp databases: Conference and Workshop Papers
Fig. 4 Number of publication in IEEE database published in 2019: Journal, Conference and
Workshop Papers
Fig. 5 Number of publication in dblp database published in 2019: Journal, Conference and
Workshop Papers
A Brief Survey on Internet of Things (IoT) 343
6 Conclusion
Internet of Things present one of the trends in the field of new technologies and
innovation, which have a big behoove in research. It has several advantages for its
users (companies and civil users), despite these benefits many challenges are present
which leads to think about developing of a new solutions to solve these. The novelty
that IoT present is the use of internet and their protocols for daily objects. In this
article we have presented a brief study of Internet of Things, its fields of applications
and issues and a brief view on the number of research works, which will prove that
IoT is an area of great importance and multidisciplinary. In the future work we wish
to develop some internet of things applications/solutions and choice an issue of this
field in the aim to propose a new solution.
References
14. Ungurean I, Gaitan N, Gaitan VG (2014) An IoT architecture for things from industrial envi-
ronment. In: 2014 10th international conference on communications (COMM). IEEE, pp
1–4
15. Nitti M, Pilloni V, Giusto D et al (2017) IoT Architecture for a sustainable tourism application
in a smart city environment. Mob Inf Syst 2017:1–9
16. Agarwal R, Fernandez DG, Elsaleh T, et al (2016) Unified IoT ontology to enable interop-
erability and federation of testbeds. In: 2016 IEEE 3rd world forum on Internet of Things
(WF-IoT). IEEE, pp 70–75
17. Ray PP (2016) A survey of IoT cloud platforms. Future Comput Inform J 1(1–2):35–46
18. Pizzolli D, Cossu G, Santoro D, et al (2016) Cloud4iot: a heterogeneous, distributed and auto-
nomic cloud platform for the IoT. In: 2016 IEEE international conference on cloud computing
technology and science (CloudCom). IEEE, pp 476–479
19. Zhou J, Leppanen T, Harjula E, et al (2013) Cloudthings: a common architecture for inte-
grating the Internet of Things with cloud computing. In: Proceedings of the 2013 IEEE 17th
international conference on computer supported cooperative work in design (CSCWD). IEEE,
pp 651–657
20. Stergiou C, Psannis KE, Kim B et al (2018) Secure integration of IoT and cloud computing.
Future Gener Comput Syst 78:964–975
21. Fagroud F, Ben Lahmar E, El Filali S (2019) Internet Of Things: statistical study on research
evolution. Int J Adv Electron Comput Sci 6(5)
22. Arnaert M, Bertrand Y, Boudaoud K (2016) Modeling vulnerable Internet of Things on
SHODAN and CENSYS: an ontology for cyber security. In: Proceedings of the tenth inter-
national conference on emerging security information, systems and technologies (SECURE-
WARE 2016), Nice, France, pp 24–28
23. Aldowah H, Rehman SU, Ghazal S, et al (2017) Internet of Things in higher education: a study
on future learning. In: Journal of physics: conference series, p 012017
24. Xiao L, Wan X, Lu X et al (2018) IoT security techniques based on machine learning: how do
IoT devices use AI to enhance security? IEEE Signal Process Mag 35(5):41–49
25. Yasuda S, Miyazaki S (2017) Fatigue crack detection system based on IoT and statistical
analysis. Procedia CIRP 61:785–789
26. Prajapati H, Mishra AK (2018) A review on baggage tracing and handling system using sensor
networks and IoT
27. Fraga-Lamas P, Fernández-Caramés TM, Suárez-Albela M, Castedo L, González-López M
(2016) A review on Internet of Things for defense and public safety. Sensors 16(10):1644
28. Suresh P, Daniel JV, Parthasarathy V, et al (2014) A state of the art review on the Internet of
Things (IoT) history, technology and fields of deployment. In: 2014 international conference
on science engineering and management research (ICSEMR). IEEE, pp 1–8
29. Alansari Z, Soomro S, Belgaum MR, et al (2018) The rise of Internet of Things (IoT) in big
healthcare data: review and open research issues. In: Progress in advanced computing and
intelligent engineering. Springer, Singapore, pp 675–685
30. Anggorojati B, Prasad R (2018) Securing communication in the IoT-based health care systems.
Jurnal Ilmu Komputer dan Informasi 11(1):1–9
31. Banerjee M, Lee J, Choo KKR (2018) A blockchain future for Internet of Things security: a
position paper. Digit Commun Netw 4(3):149–160
32. Burhanuddin MA, Mohammed AAJ, Ismail R, Basiron H (2017) Internet of Things architecture:
current challenges and future direction of research. Int J Appl Eng Res 12(21):11055–11061
33. Zhou C, Damiano N, Whisner B, Reyes M (2017) Industrial Internet of Things: (IIoT)
applications in underground coal mines. Min Eng 69(12):50
34. Asim M, Iqbal W (2016) Iot operating systems and security challenges. Int J Comput Sci Inf
Secur 14(7):314
Design and Implementation of Smart
Irrigation System Based on the IoT
Architecture
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 345
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_32
346 A. Hafian et al.
and deployment flexibility. Furthermore, the wireless sensor network possesses low
cost and highly reliable sensors, and have a transmission radius, the latter measure
autonomously the humidity, temperature parameters and communicate via wireless
technology with a central data collection station. This allows the farmer to use this
technology that provides the opportunity for producers to access remotely via wire-
less radio communication to see the field conditions and the irrigation operation, at
home or the office.
The search for an innovative and adequate material architecture (cost and perfor-
mance) for the achievement of the objectives of our work led us to do a biblio-
graphic search at the level of existing systems and related fields. For example, [3]
which presented a network of wireless sensors controlled by the broker. The latter
is in constant communication with an online server that uses the Message Queuing
Telemetry Transport (MQTT) protocol via a Wi-Fi connection that communicates
with the broker using a peer-to-peer LoRa connection. Another interesting work is
proposed by [4], which presented the fundamental design and implementation of a
wireless sensor network featuring a Zigbee protocol. This work describes the use
of a high transmission power with a relatively low energy consumption based on
a Zigbee wireless sensor network for the monitoring system of the water irrigation
control.
Also, [5] represents the design and implementation of an intelligent irrigation
system based on LoRa technology. This article presents the energy consumption
results, which take into account the different operating modes of the terminal nodes.
In this sense, to ensure adequate access to digital agriculture, we have proposed an
intelligent irrigation system with a diverse architecture using ZigBee, LoRa, and Wi-
Fi protocols based on sensor networks and wireless connectivity to all levels of the
system. The use of these different protocols in our proposed architecture allowed us
to take advantage of the different types of communication such as WLAN (wireless
local area network), WPAN (wireless personal area network), LPWAN (Low Power
Wide Area Network), which offers the possibility of remote access and at any time
to our irrigation field.
The proposed architecture far exceeds existing architectures in terms of low cost
and sophisticated performance. Indeed, it is an architecture making it possible to
benefit, on the one hand, from IoT and Cloud Computing and, on the other hand,
from hardware or software evolutions at the level of each part of the system.
Figure 1 represents the proposed system architecture which consists of five parts:
IoT nodes, local gateways, main gateway, Cloud server, and application.
Design and Implementation of Smart Irrigation System … 347
IoT Nodes are responsible for collecting data from the sensor layer components.
They are responsible for exchanging information with the local gateway, and they
receive remote irrigation commands from the gateways or the cloud. IoT Nodes send
the information obtained via the ZigBee protocol. This information from the various
sensors will be processed locally by a local gateway.
Subsequently, the main gateway performs the essential functions of collecting
data from local gateways via the LoRa protocol. Also, third-party services such as
weather forecasting can be used when deciding on irrigation programs, as well as
programming, monitoring, and diagnosis of the main gateway [6].
Using Message Queuing Telemetry Transport (MQTT), the main gateway controls
the network, which is in constant communication with the Cloud via a Wi-Fi connec-
tion. This communication is responsible for the exchange of messages between the
Cloud and the multiple local gateways of the network, which is, in turn, communi-
cate with the broker via the Lora connection. The cloud server is mainly responsible
for storing data [15] and implementing the MQTT interface for the LoRa server and
applications.
For various applications can be offered: mobile applications that developed on
both Android or iOS platforms or web applications. These applications allow users
to get the status of IoT nodes as they can also control and supervise the irrigation
system by sending commands.
Devices in the IoT network use different communication protocols with each other.
The wireless communication protocol is a set of rules that enable different electronic
devices to communicate with each other wirelessly.
The communication standard chosen to have an innovative and adequate hardware
architecture in terms of cost and performance for our proposed system are three
protocols: ZigBee, LoRa, Wi-Fi.
This connectivity meets the following constraints:
• Wi-Fi (Wireless Fidelity): Local connectivity without taking into account energy
consumption (WLAN). It is the common popular IoT communication protocol for
wireless local area networks (WLAN), which uses the IEEE 802.11 standard [14]
via 2.4 GHz UHF and 5 GHz ISM frequencies. Wi-Fi provides Internet access
to devices located approximately 20 to 40 m from the source. It has a data rate
of 600 Mbps maximum, depending on the frequency of the channel used and the
number of antennas,
• ZigBee: Local connectivity with consideration of energy consumption. ZigBee
is a wireless standard based on IEEE 802.15.4 [7]. It is a technology for robust
Wireless Personal Area Network (WPAN) communication. This technology is
now used in many equipment, very generally embedded, which imposes a very
low consumption, and are enough of a very low data rate and it ranges up to
348 A. Hafian et al.
Table 1 Characteristics of
Feature ZigBee LoRa Wi-Fi
different wireless
communication protocols Data rate [kbps] 250 110 11 * 102
Range [m] 10–100 2000 1–100
Frequency [GHz] 2.4 0.868 2.4
Power consumption [mA] 1–10 1–10 100–350
100 m [8]. The Zigbee protocol has a large number of applications in machine
to machine & IoT technologies. It offers a range of approximately 10 to 100 m
maximum, and its data transfer rate is approximately 250 Kbps,
• LoRa: Remote connectivity with low energy consumption (LPWAN). LoRa wire-
less technology is a long-range, low-power platform that uses unlicensed radio
spectrum [3] widely used in IoT applications. Low energy consumption, it offers
LPWAN connectivity between wireless sensors and the Cloud [9]. To ensure reli-
able monitoring, LoRa-based sensors can be installed in smaller devices. It aims to
eliminate repeaters, reduce the cost of devices, increase the battery life of devices,
improve network capacity and support a large number of devices [3].
Table 1 presents the characteristics of different wireless communication protocols
used in our proposed architecture [3].
3 System Implementation
be powered via the USB connection or with an external power supply. The power
source is selected automatically [10].
The expected output is then used to send the control signal via serial commu-
nication to Arduino to control the water pump to water the field accordingly. This
action can be commanded via the local gateway or remotely via applications. For the
wireless communication between IoT nodes and the local gateway, we used two of
XBee S2C Pro, which is using a 2.4 GHz wireless transceiver to communicate with
another XBee module.
Each local gateway is composed of two main elements: ESP32 (dual-core 32-
bit MCU with built-in Wi-Fi and BLE), with LoRa node chip SX1278; The LoRa
radio transceiver module capable of creating a multipoint network, with addresses
of individual nodes and encryption, within a range of about 2000 m [3].
About the main gateway we chose Raspberry Pi; it is an open-source, Linux-
based, credit card-sized computer board. It has USB ports to which you can connect
a keyboard and mouse, and a High-Definition Multimedia Interface (HDMI) video
output to which you can connect a TV or monitor. The Raspberry Pi is based on a
Broadcom BCM2835 chip. It does not feature a built-in hard disk or solid-state drive,
instead relying on an SD card for booting and long-term storage [11].
The main gateway uses the Wi-Fi connection to collect data via the LoRa protocol.
This layer also ensures that the collected signals are filtered and are ready to perform
the physical unit conversion.
For the cloud, we linked the raspberry to an MQTT broker, which will allow us
to control the irrigation system remotely. MQTT is a common protocol used in IoT
systems to connect low-level devices and sensors [12]. MQTT is used to pass short
messages to and from a broker.
The test of the different configurations was done through a miniaturized prototype
similar in terms of components to an agricultural gateway of real dimensions (Fig. 2).
We chose a site with an irrigation system to test our proposed solution. The site
is located at the Dhar ElMahraz Faculty of Sciences in Fez, Morocco (Fig. 3). The
miniaturized prototype comprises a network of 2 sensor nodes, plus the main gateway.
Every 5 min, the sensor data is acquired in the local gateway via the XBee protocol
(Fig. 4). This data will then be sent to the main gateway via the LoRa protocol.
Depending on the water needs of the crop and the climatic conditions, the main
gateway decides to act on the local gateway to operate the pump.
The loss of data is low in the case of the mobility of the nodes. The number of
retransmissions is similarly low.
For remote monitoring (application), we used the LabVIEW software to control
and supervise our system remotely, with a simplified operator interface dialoguing
via TCP/IP (Transmission Control Protocol/Internet Protocol) serial digital commu-
nication with the broker. TCP/IP is used for the transfer of data on the Internet, it
Design and Implementation of Smart Irrigation System … 351
represents in a certain way the set of communication rules on the Internet and is based
on the notion of IP addressing, that is to say, the fact of providing an IP address to
each machine of the network to be able to convey data packets.
As Fig. 5 shows, the monitoring system was used with a LabVIEW interface; it
represents the curves for soil moisture and temperature monitoring as well as the
unit of the pump.
352 A. Hafian et al.
References
1. Belkacem I (2014) L’irrigation intelligente. In: 15th annual global development conference,
Ghana
2. Kim Y, William M (2008) Remote sensing and control of an irrigation system using a distributed
wireless sensor network. IEEE Trans Instrum Meas 57(7):1379–1387
3. Gloria A, Dionisio C, Simoes G, Sebastiao P, Souto N (2019) WSN application for sustainable
water management in irrigation systems. In: 2019 IEEE 5th world forum on Internet of Things
(WF-IoT), Limerick, Ireland, pp 833–836
4. Rasin Z, Hamzah H, Mohd Aras MS (2009) Application and evaluation of high power
Zigbee based wireless sensor network in water irrigation control monitoring system. In: IEEE
symposium on industrial electronics applications, ISIEA 2009, vol 2
5. Zhao W, Lin S, Han J, Xu R, Hou L (2017) Design and implementation of smart irrigation
system based on LoRa. In: 2017 IEEE Globecom workshops (GC Wkshps), Singapore, pp 1–6
6. Lamas PF, Echarri MC, Azpilicueta L, Iturri PL, Falcone F, Caramés T (2020) Design and
empirical validation of a LoRaWAN IoT smart irrigation system. In: Proceedings, vol 42, p 62
7. IEEE 802.15.4TM (2011) IEEE standard for local and metropolitan area networks. Part 15.4:
low-rate wireless personnal area networks (LR-WPANs)
8. Jackson F, Ferial V, Jiamin P, Yan Xin P, Thierry V (2013) ZigBee, de la théorie à la pratique :
création d’un réseau ZigBee avec transmission de données. 3EI 71:1–18
9. Ayaz M, Ammad-Uddin M, Sharif Z, Mansour A, Aggoune EM (2019) Internet-of-Things
(IoT)-based smart agriculture: toward making the fields talk. IEEE Access 7:129551–129583
10. Arduino store. https://store.arduino.cc/
11. Kumar Sahu C, Behera P (2015) A low-cost smart irrigation control system. In: IEEE sponsored
2nd international conference on electronics and communication system
12. Kodali RK, Kuthada MS, Yogi Borra YK (2018) LoRa based smart irrigation system. In:
4th international conference on computing communication and automation (ICCCA), Greater
Noida, India, pp 1–5
13. Hafian A, Benbrahim M, Kabbaj MN, Bouazi A, Abouabdillah A (2020) A new approach for
optimal sizing of photovoltaic pumping systems for irrigation. In: 5th international conference
on renewable energies for developing countries, Marrakech, Morocco, pp 1–6
354 A. Hafian et al.
14. Chong C, Kumar S (2003) Sensor networks: evolution, opportunities and challenges. Proc IEEE
91(8):1247–1256
15. Roopaei M, Rad P, Choo K (2017) Cloud of things in smart agriculture: intelligent irrigation
monitoring by thermal imaging. IEEE Cloud Comput 4:10–15
Lightweight Key Exchange Algorithms
for MQTT Protocol in Its Environment
Abstract One of the most important challenges in the use of the Message Queue
Telemetry Transport (MQTT) protocol in the Internet of Thing (IoT) environment is
data confidentiality and user authentication. Moreover, it is easy to manipulate the
default security provided by MQTT during user authentication through username
and password, and it does not provide any type of encryption of the data to ensure
confidentiality or integrity. In this paper, we propose a key exchange algorithm to
provide a secret key that can be used for data encryption using permuted congruential
generator PCG-Rand as well as to ensure data integrity using Spongent lightweight
hash functions.
1 Introduction
In the last few centuries, the world has witnessed significant advancement in terms
of technology and communication. An individual’s daily life has become dependent
on the Internet of Thing (IoT), which includes most of the things or objects that are
connected to each other remotely. In IoT environments, everything becomes smarter
and is able to process data, manage and communicate with the other surrounding envi-
ronments. Moreover, it provides many services and has created a significant impact
in several fields, such as industry, smart homes, smart cities, healthcare, etc. All these
different fields collect various types of data and use a variety of techniques and proto-
cols. Currently, IoT security is one of the most popular topics in the field of research.
IoT security may include a number of different aspects, such as how to provide
confidentiality and maintain data integrity. Moreover, it provides authentication and
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 355
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_33
356 A. D. Alhejaili and O. H. Alhazmi
authorization to access smart devices, which are increasingly becoming a part of our
everyday lives. However, there are different types of protocols that can be adapted
to the IoT environment, such as Advanced Message Queuing Protocol (AMQP),
Constrained Application Protocol (CoAP), Message Queue Telemetry Transport
(MQTT), and Data Distribution Service (DDS). Each of these protocols has its own
advantages and disadvantages, and the most commonly used protocol is the MQTT
protocol.
The Message Queue Telemetry Transport (MQTT) is one of the application layer
protocols that works on Transmission Control Protocol/Internet Protocol (TCP/IP),
and it is very stable for IoT devices with limitation. According to [1] MQTT is
extremely suitable for communication that contains various types of data-link layer
protocols such as Wi-Fi. MQTT is a lightweight protocol for IoT devices, and it is
known as a publisher/subscriber protocol. The MQTT protocol contains three main
layers: the Physical layer, the TCP/IP stacks, and the MQTT application layer [2].
Moreover, it contains the following components: the publisher, the subscribers, and
the MQTT Broker (or MQTT server). Every IoT device can be a publisher; it can
connect with the MQTT Broker and publish data on a specific topic. Every subscriber
in this topic will receive the message “publish by publisher”. Figure 1 illustrates a
simple scenario to explain the MQTT publisher/subscriber concept.
Indeed, an increase in the amount of data generated by IoT devices takes place
every day, and it then becomes big data. This data is prone to various types of
attacks, which aim to obtain user information, breach privacy, and lead to unwanted
communication from the nodes due to spamming [3] and lead to Distributed Denial
of Service (DDoS). According to Vaccari et al. in [4] the SlowITe attack, which
is a type of denial of service (DoS) attack, targets the MQTT protocol as well as
all protocols that can work over TCP. Furthermore, the MQTT protocol provides
the service of user authentication through username and password, and it does not
provide encryption to ensure data confidentiality. This paper focuses on the MQTT
protocol because it is the most popular one. The contribution of this paper is the
proposed lightweight algorithm that can provide a secret key between the publisher
Subscribers
Publisher
MQTT Broker
Subscribers
and subscriber over an insecure channel based on the MQTT protocol in order to
ensure and achieve data confidentiality and authentication.
The rest of this paper is organized as follows: Sect. 2 mentions some related work
on MQTT lightweight algorithms. Section 3 provides an overview of the exploited
technologies along with a discussion of the main concepts and technologies that we
used. Section 4 discusses the proposed key exchange algorithm. The last section
includes the conclusion and a discussion on future research.
2 Related Work
MQTT protocol is one of the most popular lightweight protocols that is used in IoT
devices. The main drawback of the MQTT protocol is that it does offer security func-
tions by default and it is based on SSL/TLS, which uses username and password for
authentication. Nowadays, there are various types of attacks that target the MQTT
protocols, such as the DoS and the SlowITe attack. This creates the need for robust
security algorithms to overcome most types of attacks. Indeed, there is a various
methods that can be used to make the MQTT protocol more secure, such as encryp-
tion/decryption (by symmetric or asymmetric techniques) or digital signatures. In
the following section, we have discussed certain resources and studies that proposed
a different way of fulfilling this objective.
Symmetric block cipher or secret key encryption is one of the techniques that can
be used to create lightweight algorithms. One of the most suitable encryption algo-
rithms for the publisher/subscriber concept is Attribute-Based Encryption (ABE); it
provides access control over the data and confidentiality. The authors in [5] proposed
that the hybrid security algorithm passes through two main phases: The first phase is
the setup phase where the key authority generates a public key and a master key by
running the Attribute Based Encryption (KP-ABE) algorithm, and it distributes the
public key among all users. The second phase is the encryption where the publisher
encrypts the message by using the Dynamic S-box AES algorithm, followed by
encryption by KP-ABE with both the S-box AES Dynamic and public key as input.
As a result, the proposed algorithm provides confidentiality and access control over
data. Additionally, the study by Iyer et al. [6] proposed the use of a symmetric cipher
with 64 bits and Spongnt hash function for achieving integrity and authenticity. The
authors tried different types of lightweight block cipher, such as Speck, Present, and
Simon; the best result was obtained by using the Speck block cipher. Another study
where Bashir and Hussain in [7] proposed an algorithm to prevent unauthorized users
from accessing the data. This algorithm is based on the following: generating ID for
each node, generating a set of random numbers as block seed, generate a random
integer number used as block selector, and generating a random integer for choosing
a block, and then using all the previous numbers with the ID for encryption through
the chosen random block. Additionally, the author in [8] proposed MQTTAuth as a
security solution for the MQTT protocol. It was based on the augmented password-
authenticated key exchange (AugPAKE) algorithm, which creates a secure session
358 A. D. Alhejaili and O. H. Alhazmi
key to provide authorization and ensure the confidentiality of the transmitted data.
On the other hand, in the case of asymmetric encryption, the Elliptic-curve cryptog-
raphy (ECC) algorithm is the most popular and efficient for IoT devices. Amnalou
et al. in [9] proposed a lightweight algorithm based on the use of Elliptic Curve Diffie
Hellman Ephemeral (ECDHE) and a pre-shared key.
The relevant previous work may offer methods that show better performance,
but they consume the power of the devices due to the computational cost or make
it difficult to adapt them to all devices. Therefore, we proposed a lightweight key
exchange algorithm that can be used with the different types of a cryptography
algorithm to make it suitable for IoT devices.
This section discusses the main technologies used in the proposed algorithms.
There are various types of key exchange algorithms, such as Diffie-Hellman, Elliptic
curve Diffie-Hellman (ECDH), and ElGamal encryption. All previous of algorithm
serve the same purpose, i.e., to exchange a secret key over an insecure channel.
However, the benefit of using an elliptic curve algorithm is that it can be used to mini-
mize the length of the key. The secret key can be used with various types of encryption.
According to some studies that make a comparison between Diffie-Hellman (DH),
Rivest–Shamir–Adleman (RSA) and ECDH [10], the ECDH algorithm showed the
best result.
Spongent is lightweight hash function. It has different sizes of input (such as 88, 128,
160, 224, and 256 bits). The sponge is composed of three stages. The first stage is
the initialization where the message is padded by 0 as the initial value and cut into
the r block. The second stage is the absorbing, and it implements the XOR operation
with the first r-bit and the permutation of b and so on to the last r block in the data.
The third stage is the squeezing, and it produces the output in bits based on the user’s
choice [11].
Lightweight Key Exchange Algorithms for MQTT Protocol in Its Environment 359
The PCG stands for permuted congruential generator. It is an algorithm that is used to
generate random numbers in cryptographic algorithms. The PCG family for random
number generation is simple, easy to use, statistically efficient, and fast. Additionally,
it has a high rate of predictability of difficulty and a small code size. According to
O’Neill [12], the PCG-Rand algorithm is applicable to a range of 64 and 128-bit
streams that can provides 32-and 64-bit outputs.
The following sections discuss the proposed generate/exchange secret key algo-
rithm. Our algorithm is related to the Elliptic-curve Diffie–Hellman (ECDH) concept
but with a reduced overhead of computation. It uses PCG Rand random number
and lightweight Spongent hash (ESK-PCG-Spongent) to ensure data integrity. Our
proposed algorithms are based on the assumption of the trusted system. The trusted
system can be a secure bootstrapping that generates a prime number (PN) for
publishers and subscriber, after making sure they are authorized to be enter according
to their ID. Moreover, the trusted system is trusted by every node in the environment.
However, to ensure security, the PN will be generated for each new session. The PN
is same for the publisher and the subscriber for every session.
In this stage and after the trusted system ensure publisher and subscriber are
authorized to access, will generate PN as bootstrapping.
1. Send the prime number that generated by trusted system to MQTT broker to be
distributed on publisher and subscriber.
2. The publisher and subscriber will have their own private key (PKp and PKs,
respectively). The private key is generated by using PCG Rand random number.
3. Calculate the shared key for the publisher SHp = (PKp * PN) and the key for
the subscriber, SHs = (PKs * PN).
4. Then, after exchanging the key, the secret key will be calculated, which will be
same value for both. The secret key for the publisher will be S = (SHs * PKp)
and that for subscriber S = (SHp * PKS).
All these symbols are summarized in Table 1. Figure 2 illustrates all these steps.
360 A. D. Alhejaili and O. H. Alhazmi
Now, after generating and exchanging the secret key. The publisher and subscriber
connect with Broker through CONNECT and get CONNECTACK as response, and
subscriber subscribes to specific topic as illustrated in Fig. 2. We will use the secret
key to encrypt the MQTT payload. Then we will have the following scenario:
• Using the simple logical operation XOR to reduce the overhead and power
consumption. Also, it will use the Spongent hash to calculate the hash value
of the payload before encrypting for payload integrity. Figure 3. Illustrates XOR
encryption scenario.
In this section will discuss some attack may effect on the proposed algorithms.
• Brute-force Attack: In this type of attack the attacker exploits the length of a key
and tries to find all possible values to find the key. We assume the secret key size
is equal to 128 bit. The number of possible values to get the secret key will be as
illustrates in Eq. (1).
In addition, the average of possible number of tries to get the correct key will be
less than or equal the possible number as illustrated in Eq. (2).
ID
ID ACK
ID
ID ACK
Prime number: PN
PN PN
Publisher Subscriber
shared key shared key
Subscriber Publisher
shared key shared key
CONNE
CONNE
CONNECTAC
CONNECTAC
Subscriber
SUBACK
5 Discussion
This section will mention the improvements that have been added by our work into
the key exchange algorithm that can be used on MQTT protocol.
In our algorithm, the trusted system is the responsible for verifying the autho-
rization. The trusted system will perform the authorization using the publisher and
subscriber ID’s. Also, we reduce the overhead computation on the algorithm by using
one prime number for publisher and subscriber, which then sent out to MQTT broker
to distribute between publisher and subscriber. That means, we make the MQTT
broker is responsible for the process of distributing and exchanging keys, because
we might have more than MQTT broker on the system. Also, we use key length 128
bit, which takes a long time to be break. In addition, to ensure protection even in case
the prime number is discovered, it will be generated for every session. Furthermore,
we use a Spongent hash for the payload before encrypting it for integrity to be sent
with the payload encryption.
6 Conclusion
In the IoT environment, one of the most important factors is the security of the data.
There are various types of protocols that can be used in the IoT environment. The
MQTT protocol the most popular protocol for several reasons, and it follows the
publisher/subscriber concept. In this paper, we proposed an algorithm for creating a
secret key over an insecure channel. The proposed algorithm provides a secret key
by length 128 bit that can be used with various types of encryption algorithms. Addi-
tionally, we encrypt the payload by using XOR operation. For provide integrity, we
proposed the use of Spongent lightweight hash functions. Also, in security analysis
for the algorithm, we implemented brute-force attack on the key length and it take
many years to break it. Finally we mention to the most important improvements that
have been proposed on the algorithm.
In future work, we will implement the algorithm in the IoT environment to measure
its performance on the batteries of devices. Furthermore, we will test the performance
during different attack scenarios, such as MITM, spoofing attack, and SlowITe at-
tack. We will replace the XOR operation Scenario with other lightweight encryptions,
such as the HIGHT block cipher and the PRESENT block cipher to identify the best
result.
References
2. Vaccari I, Maurizio A, Cambiaso E (2020) SlowITe, a novel denial of service attack affecting
MQTT. Sensors 20(10):2932
3. Ajmal M, Bag S, Tabassum S, Hao F (2017) Privy: privacy preserving collaboration across
multiple service providers to combat telecoms spam. IEEE Trans Emerg Top Comput 8:313–
327
4. Vaccari I, Aiello M, Cambiaso E (2020) SlowITe, a novel denial of service attack affecting
MQTT. Sensors 20(10):2932
5. Bisne L, Parmar M (April 2017) Composite secure MQTT for Internet of Things using ABE
and dynamic S-box AES. In: 2017 innovations in power and advanced computing technologies
(i-PACT), pp 1–5. IEEE
6. Iyer S, Bansod GV, Naidu P, Garg S (December 2018) Implementation and evaluation of
lightweight ciphers in MQTT environment. In: 2018 international conference on electrical,
electronics, communication, computer, and optimization techniques
7. Bashir A, Hussain Mir A (2018) Securing communication in MQTT enabled Internet of Things
with lightweight security protocol. EAI Endorsed Trans Internet Things 3(12):154390
8. Calabretta M, Pecori R, Vecchio M, Veltri L (2018) MQTT-Auth: a token-based solution
to endow MQTT with authentication and authorization capabilities. J Commun Softw Syst
14:320–331
9. Amnalou S, Bakar KAA (2019) Lightweight security mechanism over MQTT protocol for IoT
devices
10. Goyal TK, Sahula V (September 2016) Lightweight security algorithm for low power IoT
devices. In: 2016 international conference on advances in computing, communications and
informatics (ICACCI), pp 1725–1729. IEEE
11. Bogdanov A, Knežević M, Leander G, Toz D, Varıcı K, Verbauwhede I (September 2011)
SPONGENT: a lightweight hash function. In: International workshop on cryptographic
hardware and embedded systems. Springer, Heidelberg, pp 312–325
12. O’Neill ME (2014) PCG: a family of simple fast space-efficient statistically good algorithms
for random number generation. ACM Trans Math Softw
13. Hong D, Sung J, Hong S, Lim J, Lee S, Koo B-S, Lee C, Chang D, Lee J, Jeong K, Kim H, Kim J,
Chee S (October 2006) HIGHT: a new block cipher suitable for low-resource device. In: Inter-
national workshop on cryptographic hardware and embedded systems. Springer, Heidelberg,
pp 46–59
IoT Design and Realization
of a Supervision Device for Photovoltaic
Panels Using an Approach Based
on Radiofrequency Technology
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 365
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_34
366 Z. Didi and I. El Azami
The hardware design uses an Arduino Uno board and esp32 microcontroller, a voltage
divider to measure the voltage across the photovoltaic panel and an ACS712-30A
sensor to measure the current, a 433 MHz (FS1000A) RF transmitter and the peer
receiver to perform the radio frequency communication, at the end two 16 × 02 LCD
displays for the transmitter and receiver.
Measuring electrical current and voltage and very useful for calculating energy and
power in electrical systems. In this paper, we used the current sensor ACS712-
30A (see Fig. 1) which is based on the Hall Effect magnetic field to measure the
electric current produced by the PVs, this sensor produces at output a direct voltage
proportional to the current at a rate of 0.066 V/A [8].
As you can see, this sensor is absolutely simple and contains only a few parts,
including the ASC712, passive components and connectors.
There are three variants of the ACS712 sensor depending on the current
detection interval ±5 A, ±20 A and ±30 A [8], the output sensitivity varies as
follows (Table 1):
As mentioned before, the ASC 712 is based on the Hall effect (see Fig. 2). A
copper strip joins the pins IP+ and IP−. The generated magnetic field is detected by
the sensor to transform it into an adequate voltage.
Then we are going to make a voltage divider in order to measure the voltage at the
terminals of the photovoltaic panel so as not to damage the Esp32 microcontroller.
To do this, we will integrate a voltage divider because the voltage across the Esp32
must not exceed 5 V. We take R1 = 1 K and R2 = 12 K, the output voltage is
measured at the terminals of R2 when a voltage from the photovoltaic panel is applied
to the input of these two resistors (see Fig. 3).
Based on mesh law and then Ohm’s law with the voltage at pin34 of esp32, we can
retain the authentic value of the voltage at the terminals of the photovoltaic panel.
N.B: we can easily calculate the power P consumed as well as the energy dissipated
because we have measured the two quantities, the current (I) and the voltage (V).
In this part we are going to carry out a radio frequency communication of 433 MHZ
to transmit the values Voltage, Current, Energy and Power previously determined.
We integrate the FS1000A transmitter with an Esp32 card and the proprietary
receiver with an Arduino Uno card.
The RF Receiver
The receiver associated with the FS1000A transmitter has four pins: Vcc, DATA and
the two middle pins are attached together, and the GND, (see Fig. 6).
Unlike the transmitter, the Vcc pin, the receiver must be powered by a regulated
5 V supply. GND is a ground pin (0 V).
Wiring the receiver is as easy as the transmitter. Noted that the output pin of the
receiver should be connected to pin 11 of the Arduino (see Fig. 7).
Organization Charts
• Radiofrequency transmitter.
Figure 8 shows the block diagram of the FS1000A transmitter.
• Radiofrequency Receiver
Fig. 8 Schematic
diagram—RF transmitter
Fig. 9 Schematic
diagram—RF receiver
Then we generate an ASK object called “rf_driver”. In the loop, we define a buffer
for the size of the transmitted message. You have to normalize this parameter to match
the message size, you have to include spaces and punctuation marks because they
evaluate as characters. We then examine whether we have obtained a valid package.
Finally, we use the strtok() function to capture the values: Current, Voltage, Power
and Energy [11].
Figures 11, 12, 13 and 14 show the real-time results of our project following a
433 MHZ radio frequency link to transmit the two measured and calculated quantities
IoT, Design and Realization of a Supervision Device for Photovoltaic Panels … 371
(Voltage, Current, Power and Energy) while using the FS1000A transmitter with a
card Esp32 and the receiver associated with an Arduino Uno board. It should be
noted that in this experiment we chose the simplest solution, we soldered an antenna
on the ANT pin, with this trick we increased the distance between the transmitter
and the receiver to 100 m. unfortunately this type of antenna is omnidirectional.
This method presents an inexpensive solution for the transmission of data in real
time based on radio frequency, particularly in isolated “no internet” environments,
so it is a realization of a remote supervision arrangement of photovoltaic panels.
The images which represent our project are given as follows, Fig. 11 represents
the box of the RF transmitter and Fig. 12 represents the box of the RF receiver.
372 Z. Didi and I. El Azami
Figure 13 shows the data captured on the Arduino Uno serial monitor at the
receiver following a radio frequency link with a length of 30 m between transmitter
and receiver.
These data represent current, voltage, power and total energy (Current, Ts, Power,
Energy/Min and Energy/T).
Code Test
We can finally display the total energy variation per unit of time (see Fig. 14) on the
free platform ThingSpeak (Open-Source application) for the Internet of Things, it is
a service carried by MathWorks allowing to store and collect data from.
IoT, Design and Realization of a Supervision Device for Photovoltaic Panels … 373
4 Conclusion
As you can see, adding a wireless link to Arduino implementations is very easy
and inexpensive to send/receive data one way, with 433 MHz transmit and receive
modules.
These RF modules are absolutely matched to the job at hand and the price is
affordable.
For an efficient and robust operation of the photovoltaic solar energy system, a
vigilance system is mandatory to identify difficulties and system failures before any
major failure.
By taking advantage of the different measured data, we can simply draw particular
data over time intervals, which leads to a commonly traceable pattern for detecting
operating failures. Unfortunately, this type of antenna is omnidirectional so there is
no point in radiating in the opposite direction and to the sides, which is not useful it
is necessary to better guide the radiation in a desired direction.
In addition, the development of the protection system in the solar photovoltaic
energy system increases the reliability of the system and ensures its proper func-
tioning. An effective monitoring system can help the user of that energy system keep
the system running smoothly for years to come.
References
2. Chen B, Yu Z, Liu K, Zheng X, Liu Y, Shi J, Spronk D, Rudd PN, Holman Z, Huang J (2019)
Grain engineering for perovskite/silicon monolithic tandem solar cells with efficiency of 25.4%.
Joule 3:177–190. https://doi.org/10.1016/j.joule.2018.10.003
3. Daliento S, Chouder A, Guerriero P, Pavan AM, Mellit A, Moeini R, Tricoli P (2017) Moni-
toring, diagnosis, and power forecasting for photovoltaic fields: a review. Int J Photoenergy.
https://doi.org/10.1155/2017/1356851
4. Chaibi Y, Malvoni M, Chouder A, Boussetta M, Salhi M (2019) Simple and efficient approach
to detect and diagnose electrical faults and partial shading in photovoltaic systems. Energy
Convers Manag 196:330–343. https://doi.org/10.1016/j.enconman.2019.05.086
5. Vestenicky M, Matuska S, Hudec R (2019) Simple method of photovoltaic panel power charac-
teristic measurement based on Arduino hardware platform. Transp Res Procedia 40:579–585.
https://doi.org/10.1016/j.trpro.2019.07.083
6. Yatchev I, Sen M, Balabozov I, Kostov I (2018) Modelling of a hall effect-based current
sensor with an open core magnetic concentrator. Sensors 18(4):1260. https://doi.org/10.3390/
s18041260
7. de Arquer Fernandez P, Fernandez MAF, Cand JLC, Arboleya PA (2021) An IoT open source
platform for photovoltaic plants supervision. Int J Electr Power Energy Syst 125:106540.
https://doi.org/10.1016/j.ijepes.2020.106540
8. Allegro Microsystem. ACS712: fully integrated, hall effect-based linear current sensor with 2.1
kVRMS voltage isolation and a low-resistance current conductor (2017), ACS712-DS, Rev.
7. Northeast Cutoff Worcester, Massachusetts 01615-0036 U.S.A, https://www.sparkfun.com/
datasheets/BreakoutBoards/0712.pdf. Accessed 25 Aug 2017
9. Smith DW (2013) Chapter 12 - radio transmitters and receivers. PIC projects and applications
using C. A project-based approach, 3rd edn., pp 129–134. https://doi.org/10.1016/B978-0-08-
097151-3.00012-X
10. Binh LN (2008) Multi-amplitude minimum shift keying modulation format for optical commu-
nications. Opt Commun 281:4245–4253. https://doi.org/10.1016/j.optcom.2008.04.041
11. Attaway S (2016) Matlab, 4th edn. (Cover date: 2017), pp 237–266. https://doi.org/10.1016/
B978-0-12-815479-3.00007-6
ESP8266 Wireless Indoor Positioning
System using Fingerprinting
and Trilateration Techniques
Abstract Location-based systems have come under the spotlight in the last decade.
They became significant with the bloom of applications that have tracking services.
Indoors, wireless technologies can be an alternative for precise positioning in small or
large areas where GPS signals are too weak and inaccurate. In this paper, we present
a Wi-Fi Received Signal Strength based system using fingerprinting and trilateration
techniques. We constructed radio- maps by collecting signals with ESP8266 modules
in the offline phase. Then, we looked for the high number of matches in real-time
by comparing collected signals to those stored in database. Simultaneously, we use
trilateration to find positions with three nearby APs in the online phase. We achieved
an accuracy of 98% by combining fingerprinting and trilateration locating techniques.
We also analyzed the error probability for position calculation at different time of
the day and demonstrated high stability of our system with a low standard deviation.
1 Introduction
The wide use of location services has increased with the growth and the popu-
larity of locating computation applications in indoor environments, such as airports,
schools, malls… etc. [1]. Consequently, tracking and positioning system industries
have emerged and developed smart locating technologies. Because of the consistent
emergence of location-based systems, [2] estimates that the general annual revenues
of location-based system will approximately reach 25$ billion by 2022. In fact, Indoor
localization offers an improvement of healthcare monitoring, network management,
security and personal navigation [3, 4]. Multiple indoor locating technologies such
as RFID [5], Wi-Fi [6] and Bluetooth [7] have been utilized for indoor locating
systems. Most of those technologies require massive hardware and infrastructure
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 377
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_35
378 S. El Abkari et al.
deployment, which leads to high complexity and cost maintenance. Even though the
main functionality of Wi-Fi is not locating service, Wi-Fi seems to be a promising
technology and potentially a counterpart of Global Positioning System (GPS) [8]
in indoor environment. Therefore, Wi-Fi localization using fingerprinting technique
has been adopted by many researchers because of its availability in most indoor
environments (no extra hardware needed).
Wi-Fi based locating system using fingerprinting is employed to overcome the
path-loss model issue which consists of variation of the received signal strength
(RSS) intensity at certain positions caused by multi-path fading effect in indoor
environments [9]. Thus, the knowledge of the received signal properties such as
the RSS distribution, maximum, minimum, median, mean and standard deviation
are important for the correct modeling of fingerprinting locating system. In fact,
the knowledge of necessary data number to achieve high accuracy and number of
necessary APs for precise location identification can be provided by studying the
nature of RSS and how it is affected by the environment.
In this paper, we present a Wi-Fi based positioning system using fingerprinting
and trilateration techniques. In the first stage which is the offline phase, we collect the
RSS data to construct radio-maps. In the second stage, we look for the high number of
matches while we calculate the position using trilateration. Results demonstrate high
positioning accuracy in presence of indoor constraints (fluctuations of RSS signals).
In this section, we discuss the existing literature about the wireless localization
systems (Table 1).
This section presents a survey of the most three prominent Radio Frequency
technologies used for indoor positioning as illustrated in table below.
• RFID: is a process of automatic identification using Radio Frequency communi-
cation between RFID readers and tags for Indoor locating. It determines the posi-
tion and the orientation of people or assets carrying the tag. It permits Real-Time
positioning, with less complex programming and a very high response time.
• Wi-Fi: is a wireless technology that works at high frequency waves to ensure
connectivity and communication between the network nodes in indoor locating
environment. It is suitable for locating in presence of multipath. Thus, there is
no need of extra hardware due to the availability and the popularity of Wi-Fi
infrastructure.
• Bluetooth: is a technology that uses the client-server architecture. By con-trolling
Bluetooth devices specifications such as communication parameters and emis-
sion power, we can enable the communication between the nodes of the network
and start the positioning process. It permits high data speeds but has mobility
limitations.
380 S. El Abkari et al.
• Received Signal Strength: this method is only possible for radio signals [14]. It
estimates the distance of an unknown tag to a reference access point by using the
attenuation of emitted signals strength. Basically, the smaller the RSS values the
farthest the tag is from the reference access point (AP).
• Time Of Arrival (TOA): It is a method based on the arrival time synchronization
of an emitted and received signal [14]. We calculate distance between the emitter
and the receiver using the following equation:
Multiple techniques (Table 3) are utilized in wireless based locating system. In this
section, we present two techniques used for Wi-Fi locating which are:
• Fingerprinting: it is a method consisting of two phases [15]:
– The Offline phase: which is also called calibration or training phase. At this
phase, we set up radio-maps by empirical measurements with the definition of a
reference node. The radio-maps are databases of radio signal characteristics such
as the SSID/ RSS and coordinates.
– The Online phase: wherein the process of locating is operational. The tag carried
by the user measures SSID and RSS values of the signals at an unknown position.
This value is compared then to those in the radio-map to find the higher number
of matches.
The main disadvantages of this technique are the time consumption in the offline
phase and its sensitivity to changes in the locating environments. Therefore, the
process of calibration must be repeated from time to time to guarantee a good
accuracy.
ESP8266 Wireless Indoor Positioning System ... 381
In this work, we used ESP8266 NodeMCU module as a locating node and as a mobile
tag. ESP8266 is Wi-Fi microchip with microcontroller capability and TCP/IP stack. It
is cost effective and highly integrated for various IoT applications. Table 4 illustrates
the hardware specifications of ESP8266 module.
Devices working under 802.11 standards such as ESP8266 use Clear Channel Assess-
ment protocol to prevent collision between two signals transmitted on the same
channel [20]. This method checks if any other SSID than the one emitting is transmit-
ting on the channel before sending data (frames). Table below shows the statistical
values of our three APs (ESP8266 modules) at two different locations of the user
tag (used for collection). One of the positions is near to APs while the second one is
far from the APs. This table shows that RSS mean values and standard deviation of
each ESP are similar at both position P1 and P2 (Table 5).
ESP8266 Wireless Indoor Positioning System ... 383
Table 6 Distance
SSID P1 = 1m P2 = 4m P3 = 8m
measurement at three
positions ESP1 1.72 2.56 1.84
ESP2 1.9 2.14 2.23
ESP3 1.51 2.97 1.96
Author in [25] presented the RedPin locating system that reached 90% of accuracy
in a room level. Author in [26] presented the FreeLoc that reached an accuracy less
than three meters. Manifold learning system was presented in [27]. It used four APs
and achieved an accuracy of 2.4 m. In paper [28], the author presented ADRIANE
system which achieved 2.5 m of accuracy. Forward Interpolation was introduced in
[29] and reached 2.82 m of accuracy with four APs while our system used three APs
and reached 0.46 m of accuracy (Table 9).
4 Conclusion
In this paper, we presented a Wi-Fi technology based indoor locating system using
fingerprinting and trilateration techniques. We implemented the positioning proce-
dure in ESP8266 modules and we carried our experiment in an 8 × 8 m area. In the
first stage which is the offline phase, we constructed RSS radio-maps by collecting
RSS fingerprints at certain positions. In the online phase, the mobile user carried an
ESP8266 module as a tag to read RSS values, the fingerprints of Wi-Fi and radio-
ESP8266 Wireless Indoor Positioning System ... 385
References
16. Gezici S (2008) A survey on wireless position estimation. Wirel Pers Commun 44(3):263–282
17. Gu Y, Lo A, Niemegeers I (2009) A survey of indoor positioning systems for wireless personal
networks. IEEE Commun Surv Tutorials 11(1):13–32
18. Zhang D, Xia F, Yang Z, Yao L, Zhao W (2010) Localization technologies for indoor human
tracking. In 5th international conference on future information technology, pp 1–6. IEEE
19. Ismail MB, Fathi A, Boud A, Nurdiana W, Ibrahim W (2008) Implementation of location
determination in a wireless local area network (WLAN) environment. In 10th international
conference on advanced communication technology, vol 2, pp 894–899
20. El Abkari S, Jilbab A, El Mhamdi J (2020) RSS-based indoor positioning using convolutional
neural network. Int J Online Biomed Eng 16(12):82–93
21. Barai S, Biswas D, Sau B (2017) Estimate distance measurement using Node MCU ESP8266
based on RSSI technique. In conference on antenna measurements & applications (CAMA),
pp 170–173. IEEE
22. ESP8266 nodemcu. https://www.nodemcu.com/. Accessed 27 Sep 2019
23. Laitinen E, Lohan ES (2015) Are all the access points necessary in WLAN-based indoor
positioning. In: international conference on location and GNSS (ICL-GNSS), pp 1–6. IEEE
24. Baala O, Caminada A (2006) WLAN-based indoor positioning system: experimental results
for stationary and tracking MS. In International Conference on Communication Technology,
pp 1–4
25. El Abkari S, Jilbab A, El Mhamdi A (2020) Wireless Indoor localization using fingerprinting
and Trilateration. J Adv Res Dyn Control Syst 12(7):2597–2602
26. Bolliger P (2008) Redpin adaptive, zero configuration indoor localization through user collabo-
ration. In Proceedings of the first ACM international workshop on Mobile entity localization
and tracking in GPS-less environments - MELT 2008, pp 55–60, San Francisco, CA, USA
27. Yang S, Dessai P, Verma M, Gerla M (2013) FreeLoc: calibration free crowdsourced indoor
localization. In Proceedings IEEE INFOCOM, pp 2481–2489, Turin, Italy
28. Sorour S, Lostanlen Y, Valaee S, Majeed K (2015) Joint indoor localization and radio-map
construction with limited deployment load. IEEE Trans Mob Comput 14(5):1031–1043
29. Ji Y, Biaz S, Pandey S, Agrawal P (2006) ARIADNE: a dynamic indoor signal map construc-
tion and localization system. In Proceedings of the 4th international conference on Mobile
systems, applications and services - MobiSys 2006, pp 151–164, Uppsala, Sweden
30. Peng L, Han J, Meng W, Liu J (2010) Research on radio map construction in indoor
WLAN positioning system. In: First International Conference on Pervasive Computing, Signal
Processing and Applications, pp. 1073–1077, Harbin, China
High Performance Predictive Control
for Permanent Magnet Synchronous
Machine Drive: FPGA-Based
Implementation
Abstract The use of PMSM motors is very important in industry today; they require
great robustness in terms of performance (speed, torque, flux, etc.). In this work we
present a new control strategy to improve the performance of the machine given its
non-linear model. This technique based on a new non-linear predictive controller,
which is based on the expansion of the Taylor series. The minimum variance of the
nonlinear predictive controller (CPNLVM) aims to minimize the speed error and the
robustness of the systems given the parametric variation. Another contribution in this
work is the real-time implementation on the FPGA card which will be connected to
an experimental Benchmark. First, the operating principle of the Predictive control
is developed, and then it is validated on Matlab & Simulink. Subsequently, an exper-
imental validation on a Benchmark based on PMSM and FPGA in order to test the
proposed technique. The results obtained (simulation and experimental) show the
validation of the proposed model and also a great improvement in terms of system
performance.
1 Introduction
B. Bossoufi (B)
LIMAS Laboratory, Faculty of Sciences Dhar El Mahraz, Sidi Mohamed Ben Abdellah
University, Fez, Morocco
A. Lagrioui
Department of Electrical and Computer Engineering The Higher National School of Arts and
Trades, Moulay Ismail University, Meknes, Morocco
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 387
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_36
388 B. Bossoufi and A. Lagrioui
improves their performance. Recently, the start-up of the PMSM without speed
sensor has attracted great interest in the industry where the use of position sensors is
limited. This advantage reduces the cost of training and improves the robustness of
the machine. The quality of the performance of the PMSM machine is mainly based
on knowledge of the position and speed of the rotor. The control strategies for esti-
mating the position of the sensorless rotor are mainly based on the non-linear PMSM
motor model. Most of the control strategies are based on the motor model detecting
the FEM, which is directly related to the position and speed of the rotor, using esti-
mation observers. Other strategies are based on flux estimation in the machine model
[2]. In addition, adaptive estimate observers were used to estimate the stator current,
speed and position of the rotor.
The predictive controller used in this work is a nonlinear technique of minimal
variance (CPNLVM) which is based on the expansion of the Taylor series; its objec-
tive is to minimize the error between the output and the reference applied to the
machine.
Currently, the use of FPGA technology has become very important in various fields
(signal processing, Telecom, Control systems, Calculators, Robotics….) thanks to
its simplicity of operation and its memory capacity and the number of ‘I/O and its
price compared to other types of cards (DSP, microcontrollers).
The field of electric machine control has known a great evolution in the industrial
field thanks to this type of technology, several techniques have been successfully
implemented on the FPGA target (PWM, FOC, Backstepping, DTC,….). In this
work we are interested in the implementation of a new control technique (predctive)
on the FPGA card in order to improve the performance of the systems in terms of
speed and robustness.
Figure 1 shows the general structure of the control of a PMSM motor:
With:
⎡Rs L sq ⎤
⎡ ⎤ ⎢ −.i + p .i .
L sd
sd
L sd
sq
⎥
f 1 (x) ⎢ ⎥
⎢ ⎥ ⎢⎢ R s L sd f ⎥
⎥
f (x) = ⎣ f 2 (x)⎦ = ⎢ − .i sq − p .i sd . − p . ⎥ (4)
⎢ L sq L sq L sq ⎥
f 3 (x) ⎣ ⎦
f 3p 3p
− . + (L sd − L sq ).i sd .i sq + f .i sq
J 2J 2J
With b represents the disturbance which influences the regulation of the outputs,
its dynamics are slow compared to the state variables. We have:
db
=0 (5)
dt
The cost function proposed to develop the predictive control technique for PMSM
is written as follows [5]:
T1
T2
1 1
= (ed (t + τ )) dτ +
2
(e (t + τ ))2 dτ (6)
2 2
0 0
390 B. Bossoufi and A. Lagrioui
ed (t + τ ) = y1r (t + τ ) − y1 (t + τ )
(7)
e (t + τ ) = y2r (t + τ ) − y2 (t + τ )
⎨ y1 (t + T1 ) y2 (t + T2 ) T = i sd (t + T1 ) (t + T2 ) T
⎩ y (t + T ) y (t + T )
T = i
T (8)
1r 1 2r 2 sdr e f (t + T1 ) r e f (t + T2 )
The objective of this control technique is to find the control quantities which
minimize the performance error defined by (7).
The exits and their future instructions are calculated approximately by the expan-
sion of the Taylor series. The error between the output yi and the reference yir is
repeated ρi time to reveal the regulation, hence [8]:
⎧
⎨ y1 (t + τ ) y2 (t + τ ) T = (τ )Y (t)
⎩ y (t + τ ) y (t + τ )
T = (τ )Y (t)
(9)
1r 2r r
With
Y (t) = [y1 y2 ẏ1 ẏ2 ÿ2 ]T
(10)
Yr (t) = [y1r y2r ẏ1r ẏ2r ÿ2r ]T
1 (τ ) 10τ 0 0
(τ ) = = (11)
1 1 0 τ τ2
2
2 (τ )
Using (9) in Eq. (6), the quadratic performance criterion will be:
1
= (Y (t) − Yr (t))T (T1 , T2 )(Y (t) − Yr (t)) (12)
2
T1
T2
(T1 , T2 ) = (1 (τ )) 1 (τ )dτ +
T
(2 (τ ))T 2 (τ )dτ (13)
0 0
T2
⎡ ⎤
T1 0 21 0 0
⎢ 2 3 ⎥
⎢ 0 T2 0 T22 T62 ⎥
⎢ T2 ⎥
(T1 , T2 ) = ⎢ T13
⎢ 2 02 3 03 04 ⎥
1 ⎥ (14)
⎢ T2 T2 T2 ⎥
⎣ 0 2 0 3 8 ⎦
T3 T4 T5
0 62 0 82 202
In the absence of disturbance, the law of control minimizing the cost function is
obtained from the resolution of (16) [9]:
d
=0 (16)
du
This leads us to solve the following equations:
T
d d
= (Y (t) − Yr (t) (T1 , T2 )(Y (t) − Yr (t)) = 0 (17)
du du
⎛⎡ ⎤⎞T ⎛⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎞
01×2 y1 (t) − y1r (t) 01×2 01×2
⎜⎢ ⎥⎟ ⎜⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎟
⎜⎢ 01×2 ⎥⎟ ⎜⎢ y2 (t) − y2r (t) ⎥ ⎢ 01×2 ⎥ ⎢ 01×2 ⎥ ⎟
⎜⎢ ⎥⎟ ⎜⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎟
⎜⎢ L g h (x) ⎥⎟ (T )⎜⎢ L f h 1 (x) − ẏ1r (t) ⎥ + ⎢ L g h (x) ⎥u(t) + ⎢ L g h (x) ⎥b(t)⎟ = 0 (18)
⎜⎢ 1 1 ⎥⎟ ⎜⎢ ⎥ ⎢ 1 1 ⎥ ⎢ 2 1 ⎥ ⎟
⎜⎢ ⎥⎟ ⎜⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎟
⎝⎣ 01×2 ⎦⎠ ⎝⎣ L f h 2 (x) − ẏ2r (t) ⎦ ⎣ 01×2 ⎦ ⎣ L g2 h 2 (x) ⎦ ⎠
L g1 L f h 2 (x) L 2f h 2 (x) − ÿ2r (t) L g1 L f h 2 (x) L g2 L f h 2 (x)
The expression developed in the predictive control law (CPNLVM) is that of the
matrix G1 (x) defined by (6). For this we define a generalized predictive command
matrix G2 (x) is given by:
392 B. Bossoufi and A. Lagrioui
− L1sd 0
L g2 h 1 (x)
G 2 (x) = = 3 p(L sd −L sq )i sq
L g2 L f h 2 (x) + K 12 L g2 h 2 (x) − 2J L sd
1
J
−K 2
1 + f
J
(21)
The dynamic error taking into account the disturbance can be written under:
K 01 ed (t) + K 11 ėd (t) = 0
(22)
K 02 e (t) + K 12 ė (t) + K 22 ë (t) = 0
From Eq. (23), the system is stable. Otherwise, the prediction time is very small
which shows a great speed of the system. This is to the detriment of torque control
[10].
In cases where the perturbation constraint is not affected, its estimation will be
used. In the same way as in the case of predictive control (CPNLVM), the equation
of the disturbance estimator is written as follows:
⎡ t
⎤
⎢ μd K 1
0 ed (τ )dτ + ed (t) ⎥
b̂(t) = ⎢
⎣
t
0 ⎥
⎦ (24)
μ K 02 e (τ )dτ + K 12 e (t) + ė (t)
0
We also show that the stability of the controller (24) is ensured by:
2
μd K1 f
< 0 ; μ − 2 <0 (25)
L sd J J
Once the observer blocks are saturated, the law of control takes the following
form:
∗
u dr e f (t) u d (t) u d (t)
u(t) = = + = u ∗ (t) + u(t) (26)
u qr e f (t) u q∗ (t) u q (t)
High Performance Predictive Control ... 393
⎛⎡ ⎤ ⎞
1
(i)
⎜⎢ i=0 K i1 y1r (t) − L if h 1 (x) ⎥ ⎟
u ∗ (t) = G 1 (x)−1 ⎜ ⎢
⎝⎣ 2 ⎥
∗ ⎟
⎦ + G 2 (x)b̂ (t)⎠ (27)
(i)
K i2 y2r (t) − L if h 2 (x)
i=0
⎛ ⎞
t
u(t) = −⎝G 1 (x)−1 G 2 (x)μ G 1 (x)(u(τ ) − sat (u(τ )))dτ ⎠ (28)
0
The limit values of the control variables are identical to the variables developed
for predictive control (CPNLVM).
The results obtained by applying a predictive observer on the PMSM, are given for
the following parameters:
• the sample time Tsc = 50 μs,
• the prediction time to control the current T1 = 0.5 ms
• prediction time to control the speed T2 = 5 ms.
• The constant of the controller: μd = –0.1 and μ = 0.00001.
In order to evaluate the performance of the predictive command applied to the
PMSM machine, we performed certain simulations on Simulink:
• Pursuit test or trajectory follow-up: to check the continuation of the speed.
• Robustness test: in order to guarantee correct operation despite the parametric
variation of the machine. The test was carried out for a change of the machine
parameters (Rs = 0,5Rsn, Lsq = 0,5Lsqn, Lsd = 0,5Lsdn, f = 0,5fn, J = 0,5Jn).
Another change in flux with an increase of 20%.
The Fig. 2 shows the general structure of the predictive controller based on the
equations developed in part 3:
400
5 10
La vitesse de rotation et sa rÈfÈrence
is1(A)
Les courants direct et quadratique
is2(A)
Les courants statoriques
300 4
wref (Rad/s) is1(A)
5
w (Rad/s) 3
200
isd(A)
2 isq(A)
100 0
1
0
0 -5
-100
-1
-10
-200 0 1 2 3 4 5 6 7 8 9 10
0 1 2 3 4 5 6 7 8 9 10 -2
0 1 2 3 4 5 6 7 8 9 10 t/Te 4
t/Te x 10
4 x 10
t/Te 4
x 10
300 12 10
Ce (N.m)
Les e coules de charge Cr
et ÈlectromagnÈtique Ce
200 10 Cr (N.m)
5
wref (Rad/s) 8
100
w (Rad/s)
6 0
0
4
-100 -5
2
-200 0
0 1 2 3 4 5 6 7 8 9 10 -10
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10
4
t/Te x 10 t/Te 4 4
x 10 t/Te x 10
The results obtained in this test (Fig. 3) show better tracking for speed with zero
dynamic error.
High Performance Predictive Control ... 395
Erreur de Vitesse
150 wref (Rad/s) 0.06
w (Rad/s)
100 0.04 1.5
isq(A)
isd(A)
50 0.02
1
0 0
0.5
-50 -0.02
-100 -0.04
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10 0
0 1 2 3 4 5 6 7 8 9 10
t/Te 4
x 10
t/Te 4
x 10 t/Te 4
x 10
0.08
2 200
0.06
Erreur de vitesse
1 150
wref (Rad/s)
w (Rad/s)
0 100 0.04
-1 50
0.02
-2 0
-3 -50 0
-4 -100
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10
-0.02
t/Te 4 t/Te x 10
4
0 1 2 3 4 5 6 7 8 9 10
x 10
t/Te 4
x 10
Fig. 5 Performance of the controlled system CPNLVM in case of a fast dynamic trajectory with
anti saturation scheme
FPGA technology (Actel, Xilinx and Altera…) offers a very interesting reconfig-
urable structure for the implementation of control strategies. It offers great flexibility
in programming and design. Currently, the FPGA offers the use of Blocks dedicated
to the field of control (PWM, PI).
The FPGA architecture was designed using CAD tools. The description of the
program is done graphically (ia blocks) or via an HDL (Hardware Description
Language) VHDL or Verilog programming language. These two languages are stan-
dardized and provide description at different levels, and above all the advantage of
being portable and compatible with all the FPGA technologies previously introduced.
Figure 5 summarizes the different steps for programming an FPGA card. A
programming tool (System Generator: merger of Xilinx and Matlab), allows you
to generate the command part program (already done on Simulink) in VHDL code
through routing steps, Timming, and I / O allocation. Thereafter the SG generates
a VHDL file to be downloaded into the memory of the FPGA. This file is called
bitstream.
396 B. Bossoufi and A. Lagrioui
In order to test the validation of the proposed control model, a test bench was
produced in the laboratory based on an FPGA card and PMSM machine (Fig. 6). For
this work, we use a FPGA card type XC3S500E Spartan3E from Xilinx.
1. Experimental results
Figure 7 presents the scheme of predictive control on Simulink-Xilink (System
Genertaor). The sampling frequency for this implementation is 20 kHz.
The functional model for predictive Controller from SYSTEM GENERATOR is
composed the different blocks:
• IC coder interface block to adapt the input of the FPGA with the position sensor.
High Performance Predictive Control ... 397
Fig. 9 Stator current, abc-axis current in the PMSM, d-axis and q-axis current in the PMSM
• ADC analog analog printer interface block to adapt the input of the FPGA and the
converter (ADCS7476MSPS 12 bits A/D) which is linked to the current sensors
(Hall effect)
• Transformation blocks of Park (abs-dq) and Park Inverse (dq-abc),
• SVW block which is the most important, to provide switching signals to the IGBTs
of the inverter in the power section.
• Predictive control block to control the speed and stator currents of the PMSM,
• “Timing” block to manage the period of use of each bloc,
• RS232 block for signal synchronization and signal acquisition.
The implementation of the predictive control on FPGA card is characterized by a
reduced operation time.
398 B. Bossoufi and A. Lagrioui
6 Conclusion
This work was dedicated to the application of a new predictive control technique with
estimators on a PMSM-based system. As well as the real-time implementation of the
control algorithm on the FPGA target. The proposed model has been successfully
validated (pursuit and robustness) by a simulation on Simulink and also by practice
on a benchmark based on FPGA.
References
1 Introduction
SLAM [1] is one of the ongoing research topics in the autonomous driving and robotic
area. However, SLAM is a heavy computational and complex algorithm that contains
compute-intensive must be optimized and implemented in suitable embedded archi-
tecture. The compute-intensive task is often managed by CPU, but it remains limited
according to characteristics and number of cores. Rather, GPU is usually used for this
purpose even though it enforces some limitations on accelerated algorithms. Thus,
Field Programmable Gate Array (FPGA) [2] is an attractive accelerator to bridge this
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 399
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_37
400 A. Mamri et al.
gap, which provides low power, low cost, high speed, and significant gains. VHDL
and Verilog are the recommended hardware description languages for FPGA, except
they are often more difficult in the case of complex algorithms. The High-Level
Synthesis (HLS) [3] is then developed to convert C/C++ programming language to
the hardware description, taking into account of low usage of FPGA resources. In
the ongoing research framework, OpenCL [4] HLS based aims for targeting a new
trend of heterogeneous FPGA-based embedded architectures. Moreover, OpenCL
is a portable language framework that provides the capability to explore multiple
embedded accelerators, including FPGAs, GPUs, and DSPs. Besides, CUDA [5]
is a high-level language developed by NVIDIA to handle its specific GPU-based
heterogeneous systems.
The ultimate goal of this work is to improve performance and embed ORB-SLAM
[6, 7], Visual Odometry (VO) [8], and Visual Inertial Odometry (VIO) through effi-
cient hardware/software optimization of a time-consuming functional block. Besides,
the proposed optimization provides a parallel implementation that handles sequential
parts using high-level OpenCL synthesis for FPGA-based heterogeneous embedded
architectures and CUDA for heterogeneous NVIDIA GPU-based architectures. The
comparative study is then conducted between the original system and the proposal
with parallel architectures and languages. The reminder of this paper is structured
as follows. The related work in Sect. 2. The targeted functional block is presented
in Sect. 3. The Sect. 4 presents the OpenCL software optimization and the hardware
implementation of the proposed optimization. The heterogenous embedded hardware
is described in Sect. 5, with results. Finally, a conclusion in Sect. 6.
2 Related Work
Recent works based on the ORB-SLAM algorithm target the feature tracking process
to deal with the computational time using heterogeneous CPU-GPU architecture.
Nguyen et al. [9] developed a novel VSLAM system named HOOFR SLAM based
on a FAST detector and amended FREAK descriptor. Afterward, they adopt the
OpenMP multithreading language to implement HOOFR extractor on CPU due to the
machine learning optimization, whereas the HOOFR descriptor is ported into GPU
accelerator using OpenCL. As a result, their proposal achieved remarkable gains
and effective speedup compared to stereo ORB-SLAM2 [10]. Jincheng Li et al. [11]
presented an improved MultiCol-SLAM using embedded heterogeneous CPU-GPU
architecture. Thus, they attempted to improve and parallelize ORB feature extraction
using CUDA and then implement it on GPU accelerator. Besides, they proposed a
CPU-based multithreading pipelining method to solve CPU and GPU load imbalance.
As a result, their proposal improved the frame rate and achieved a higher speedup and
low resource utilization than the CPU implementation. There are similar works of
ORB-SLAM2 with CPU-GPU implementation in [12, 13]. However, these works are
unfinished and require a low power consumption for embedded applications, which
is guaranteed by heterogeneous CPU-FPGA architecture.
High-Level Synthesis Implementation of Monocular SLAM ... 401
3 Algorithm Study
ORB-SLAM contains three concurrent threads; tracking, local mapping, and loop
closing. The tracking thread localizes the camera at each acquired frame and decides
to add it to the system. For the local mapping thread, the system is based on the
added keyframes to execute high-cost Bundle Adjustment (BA) [18] to reconstruct
an optimal map. At the end of the process, a loop closing thread is achieved to correct
the estimated localization based on the detected potential loops at each received
keyframe.
The proposed optimization concerns the monocular ORB SLAM2, the monoc-
ular ORB-SLAM3, VO, and VIO. The monocular ORB-SLAM2 is selected to
402 A. Mamri et al.
investigate the time-consuming blocks, then evaluated to lead toward the targeted
block. Table 1, shows the evaluation by the CPU of the targeted platforms using
TUM1 dataset [19].
The evaluation of the three threads shows that MapPoint culling and Local BA
are the most time-consuming blocks that require high dependency with other blocks.
Hence, the proposed optimization aims to parallelize the tracking thread, especially
the third time-consuming block called Initial pose estimation.
Map initialization is a part of the Initial pose estimation functional block (FB2) that
handles the monocular initialization of the relative camera pose between two frames
using two geometrical models (fundamental matrix F [20] for the non-planar scene
and a homography H [20] for the planar scene). However, it performs 30% of the
tracking thread and 67% of FB2. Practically, CPU carries out the two geometrical
models F and H using C++ multi-threading API. For the sake of accuracy, each
model count on the normalization method that handles 2001–2010 keypoints. Mean-
while, the normalization
method
(Normalize) is executed consecutively
twice for the
reference image u ri xir , yir ∈ Fr and the current image u ic xic , yic ∈ Fc into each
model. Therefore, the main idea of this optimization is to introduce the first step to
implement the initial pose estimation functional block by accelerating the iterative
normalization method in one execution available to H and F into FPGA and GPU
accelerators. Normalize function computes at first the mean xr,c , yr,c of Nr,c corner
coordinates in both current and reference frame with the following equation:
With:
1 1
x =
sr,c , y =
sr,c (4)
xr,c
dev yr,c
dev
4 Heterogeneous Implementation
This section describes the accelerated Normalize version step by step. The descrip-
tion is based on OpenCL for FPGA, whereas the CUDA version could be inferred
easily for GPU. Normalize function contains equations that are parallel in nature,
while others are not obvious to parallelize and require a special modification of the
algorithm to adjust to the FPGA kernel. To bridge this gap, we propose OpenCL
optimization for Normalize in order to parallelize the sequential parts and efficient
implementation of FPGA-based heterogeneous architecture with efficient usage of
FPGA resources.
For the sake of the proposal clarity, a good understanding of the software and hard-
ware is required. Algorithm 1, provides insight into the software, where the host
c
adjusts the input/output by resizing the current keypoints kpsx,y and the reference
r
keypoints kpsx,y to the closest required value by the architecture (i.e., 2010 to 2048)
in line 3, and then merges them in one single vector kps in line 4. It has to be noted
r c
that kpsx,y and kpsx,y are vectors combined of their coordinates (x, y). Instead of the
CUDA optimization, the coordinates of each keypoints are buffered as arguments to
the GPU without fusion.
404 A. Mamri et al.
The HLS creates more efficient hardware structures for the kernel and schedules
the buffer kps in the constant memory for a reason to process each kps coordinate
in their reserved work group. As depicted in Fig. 1, the HLS provides a hardware
description of the proposal using 4 work groups with (6) and 512 work items on each
work group with (7).
kps size
nbr work group = (6)
one coordonate of kps
kps size
nbr work item = (7)
nbr work group × nbr of vectorization
Fig. 1 The proposed OpenCL software and the data flow on each workgroup with kps (blue),
localSum (gray) and T (green)
(a ) (b )
throughput and reduces memory accesses. On the other hand, the memory concep-
tion of FPGA provides LSU embedded cache that maintains high bandwidth and
short latency. Instead of GPU, that has a heavy memory clock rate, which requires a
Direct Memory Access (DMA) system to overcome the memory bank conflict.
6 Conclusion
This work offers an efficient strategy to tackle the computational time of the compute-
intensive parts using heterogeneous architectures. Furthermore, the main idea is to
present FPGA-based heterogeneous architecture capability to handle the computa-
tional complexity and improve the performance of the ORB-SLAM system, instead
of using only GPU-based heterogeneous architecture. Thus, our further work will
adopt a high-performance scalable compute FPGA accelerator as ARRIA 10 to attain
a complete end-to-end embedded SLAM system.
References
9. Nguyen D-D, Elouardi A, Florez SAR, Bouaziz S (2018) HOOFR SLAM system: an
embedded vision SLAM algorithm and its hardware-software mapping-based intelligent
vehicles applications. IEEE Trans Intell Trans Syst 20(11):4103–4118
10. Mur-Artal R, Tardos JD (2017) ORB-SLAM2: An open-source SLAM system for monocular,
stereo, and rgb-d cameras. IEEE Transac Robotics 33(5):1255–1262. https://doi.org/10.1109/
TRO.2017.2705103
11. Li J, Deng G, Zhang W, Zhang C, Wang F, Liu Y (2020) Realization of CUDA-based real-time
multi-camera visual SLAM in embedded systems. J Real-Time Image Process 17(3):713–727.
https://doi.org/10.1007/s11554-019-00924-4
12. Aldegheri S, Bombieri N, Bloisi DD, et al (2019) Data flow orb-slam for real-time performance
on embedded gpu boards. In 2019 IEEE/RSJ International Conference on Intelligent Robots
and Systems (IROS), pp 5370–5375. IEEE
13. Bourque D (2017) Cuda-accelerated orb-slam for uavs
14. Abouzahir M, Elouardi A, Latif R, Bouaziz S, Tajer A (2018) Embedding SLAM algorithms:
has it come of age? Robot Auton Syst 100:14–26
15. Rosten E, Porter R, Drummond T (2010) Faster and better: A machine learning approach to
corner detection. IEEE Trans Pattern Anal Mach Intell 32(1):105–119. https://doi.org/10.1109/
TPAMI.2008.275
16. Shi X, Cao L, Wang D, Liu L, You G, Liu S, Wang C (2018) HERO: accelerating autonomous
robotic tasks with FPGA. In 2018 IEEE/RSJ International Conference on Intelligent Robots
and Systems, pp 7766–7772
17. Kohlbrecher S, von Stryk O, Meyer J, Klingauf U (2011) A flexible and scalable SLAM system
with full 3D motion estimation. In 2011 IEEE International Symposium on Safety, Security,
and Rescue Robotics (SSRR), pp 155–160
18. Triggs B, McLauchlan PF, Hartley RI, Fitzgibbon AW (2000) Bundle adjustment a modern
synthesis. In Vision Algorithms: Theory and Practice. New York, NY, USA, pp 298–372.
Springer
19. TUM1 dataset, https://vision.in.tum.de/data/datasets/rgbd-dataset/download
20. Hartley R, Zisserman A (2003) Multiple view geometry in computer vision. 2nd ed., Cambridge,
U.K.: Cambridge Univ. Press
21. Altera sdk for OpenCL, Best Practices Guide. https://www.intel.com/content/dam/www/pro
grammable/us/en/pdfs/literature/hb/opencl-sdk/aocl-best-practices-guide.pdf
Improving Multispectral Image
Processing for Real-Time Agricultural
Indices Tracking Using an Embedded
System
Abstract Recently, embedded systems have been used in precision agriculture for
application control, tracking, and agricultural field monitoring. The integration of
embedded systems in precision agriculture will help much better in making deci-
sion. This needs the study of such systems in terms of complex algorithms processing
and results from consistency. This study requires software optimizations and hard-
ware requirements to speed up processing times while keeping the results reli-
able. This work presents the evaluation of the most known vital indices in agricul-
ture precision algorithms on a heterogeneous embedded system using the OpenCL
parallel programming language. The algorithm evaluated in this work is dedicated
to agricultural field tracking.
Two implementations are compared: the OpenMP optimization on a multi-CPU
architecture and the OpenCL on a CPU-GPU heterogeneous architecture. In this last
case, results show improvements of 45x in speedup, enhancing image processing-
based agriculture field tracking.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 411
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_38
412 A. Saddik et al.
The evaluation of the algorithm was based on the OpenCL parallel programming
language. OpenCL represents a combination of the world of GPU, CPU, and FPGA.
414 A. Saddik et al.
camera’s strong point is its low weight (72 g) compared to the RedEdge camera (250
g). This camera offers four separate bands and one RGB image with a resolution of
4608 × 3456 pixels and the other bands with a resolution of 1280 × 960 pixels [23].
The second block is dedicated to processing images emitted by the camera. This
processing is done in real-time to follow the number of time-laps delivered each
second. This real-time processing is based on the calculation of vital indices, namely
the normalized difference vegetation index (NDVI) and the normalized difference
water index (NDWI). The third block is designed to apply the threshold operation
to classify agricultural fields and the storage of classified data. The index evaluation
was based on a heterogeneous CPU-GPU system, using both OpenCL and OpenMP
languages. Figure 2 shows the algorithm used in this work.
The complexity of this algorithm is of O = k × i × j with K the number of images
desired i and j is the size of the image. The processing of the algorithm is based on
the calculation of NDVI and NDWI as mentioned in the Eq. 1 and 2.
N I R − RE D
N DV I = (1)
N I R + RE D
G − NIR
N DV I = (2)
NIR +G
The evaluation results showed a great benefit compared to the published work
[14]. Our approach showed that we can achieve improvement with 90x of speedup.
This will allow us to monitor transfer of images collected by the cameras used for
the database. Table 1 shows the processing time obtained by different tools. The time
obtained with OpenCL has shown a great optimization of the processing time that
keeps cameras’ acquisition frequency. This optimization will also allow adding other
blocks for image pre-processing and compression afterward. The architecture of the
implementation on the XU4 board is shown in Fig. 3. After evaluating each block’s
algorithm, we concluded that the second block consumes more time than the other
block. For this reason, we opted to send the second block to the GPU for acceleration,
as shown in Fig. 3.
The following figure shows the NDVI and NDWI evaluation results. The left
image shows the NDWI values using an image collected by a ground robot, the
bluest regions are close to the value 1. Similarly, for the image on the right using a
UAV, the reddest regions have a higher NDVI index.
Figure 4 shows two images resulting from the NDVI and NDWI based assessment
with a data matrix, which helps us have a different map containing the index values.
This map allows farmers to determine which areas lack water or vegetation and make
decisions manually. In case we need an autonomous system that automatically makes
the decision, we need a thresholding operation that will separate the different regions.
We can perform thresholding techniques as performed in Otsu [24], but the use of
these techniques has increased the algorithm complexity as well as the processing
time. For this reason, we used a simple binary thresholding operation. Figure 5 shows
the NDVI and NDWI evaluation results. The middle image shows the NDWI values
using an image collected by a ground robot; the bluest regions are close to the value
1. Similarly, for the image on the right using a UAV, the reddest regions have a higher
NDVI index.
4 Discussion
The thresholding operation is very important in our case for the classification of
agricultural fields. In the image on the left of Fig. 5 we used a thresholding operation
with an index of NDVI = 0.5. With this value, we can select only regions in white
with a higher value of 0.5. These regions will be in black on the other case (NDVI
lower than 0.5). The right image is based on a threshold value of 0.3, which shows
more regions with an index of more than 0.3. In the middle image, we used the red
color to select the desired regions. The use of colors in the built maps allows us to
have a single map with different NDVI or NDWI values. The figure shows the NDVI
thresholding operation.
After the operation is proceeded by the third block, it is necessary to generate
a map with different colors. In Fig. 6 the regions that have a high index close to
Fig. 7 Evaluation of 80
images with different tools
1 will be colored red. Otherwise, the regions that have an index close to 0 will be
yellow. This operation allows the classification based on the generated map. In this
context, we can find software that allows us to do the classification. In this context,
authors in [25] used the eCognition software to classify satellite images. The limit of
these techniques is their processing time that requires offline processing and hence
causes problems to applications that need fast processing in order to give decisions
in real-time.
The logarithmic curve in Fig. 7 shows the processing time for different imple-
mentations. The green curve presents the processing time based on XU4 and C/C++
language. In this case, we have a time that varies between 80 ms and 100 ms with an
average of 91 ms, the blue curve gives the C/C++ implementation on the desktop. The
curve in purple and red show the processing times using OpenCL and OpenMP imple-
mentations. From this figure, we can conclude that heterogeneous systems, based on
CPU-GPU, have a very important impact on the processing times. In this case, an
advanced study of the Hardware-Software Co-Design approach is recommended
to propose more optimized implementations in terms of memory consumption and
execution times.
Figure 8 shows the processing time using an OpenCL implementation on the XU4
architecture. Using OpenCL allows us to exploit the GPU part (Mali-T628 MP6 @
600 MHz) of the XU4 and the CPU cores (Quadra-core ARM A57 @ 1.9 Ghz,
Quadra-core ARM A7@ 1.2 Ghz). The variation in execution time depends on the
amount of information in the images. Generally, the time varies between 1.5 ms and
0.85 ms, with an average of 1.1 ms. This meets the constraint of real-time processing
in embedded systems.
The processing time obtained shows the heterogeneous system’s strength, which
can accelerate algorithms in the GPU part. The heterogeneous programming tool
allows us to use both features of the XU4 board, despite the use of OpenMP, which
Improving Multispectral Image Processing ... 419
allows using both processors of the CPU part of the XU4 board, but it remains limited.
This limitation is shown in Table 2, which presents a processing time of 47 ms. This
time obtained is minimized compared to the time proposed in C/C++; it only gives
21 fps processing compared to C/C++, which gives just 10 fps. On the other hand,
the use of OpenCL in the GPU part of the board benefits ×45. this optimization, we
denounce an idea on the use of this type of system.
5 Conclusion
The use of embedded systems in precision agriculture has become a strong point for
decisional systems. This becomes possible thanks to the new generation of embedded
architectures’ performances, either homogeneous (CPU, GPU, FPGA) or heteroge-
neous (CPU-GPU, CPU-FPGA). In this work, we have presented an implementation
of our algorithm for tracking different vital indices on a homogeneous and hetero-
geneous architecture based embedded system. The results show that the use of the
OpenCL-based CPU-GPU implementation has a benefit of 90x of speedup than that
given in [14], and 50x of speedup tan the OpenMP implementation, which reduces
the processing time for real-time decisions. As future work, we aim to validate the
algorithm implementation on the Odroid XU4 platform using the HIL (Hardware In
the Loop) evaluation and a dataset acquired with a UAV operating in a referenced
environment.
Acknowledgment We owe a debt of gratitude to the National Centre for Scientific and Technical
Research of Morocco (CNRST) for their financial support and for their supervision (grant number:
19 UIZ2020).
References
1. Berni JAJ, Zarco-Tejada PJ, Suarez L, Fereres E (2009) Thermal and narrowband multispectral
remote sensing for vegetation monitoring from an unmanned aerial vehicle. IEEE Trans Geosci
Remote Sens 47(3):722–738
2. Baek J-Y, Jo Y-H, Kim W, Lee J-S, Jung D, Kim D-W, Nam J (2019) A new algorithm to
estimate chlorophyll-a concentrations in turbid yellow sea water using a multispectral sensor
in a low-altitude remote sensing system. Remote Sens 11:2257
3. Guan K et al (2018) Mapping paddy rice area and yields over thai binh province in viet nam from
MODIS, landsat, and ALOS-2/PALSAR-2. IEEE J Selected Topics Appl Earth Observations
Remote Sens 11(7):2238–2252
4. Sa I et al (2018) weedNet: dense semantic weed classification using multispectral images and
MAV for smart farming. IEEE Robot Autom Lett 3(1):588–595
5. Hu Z et al (2019) Application of non-orthogonal multiple access in wireless sensor networks
for smart agriculture. IEEE Access 7:87582–87592
6. Maxime M (2019) High level approach for accelerating algorithms on CPU/GPU/FPGA hetero-
geneous architectures. Application to the qualification of radars and electromagnetic listening
systems. Parallel, distributed and shared computing. Paris-Saclay University, French
420 A. Saddik et al.
7. Liu X et al (2019) Monocular camera based fruit counting and mapping with semantic data
association. IEEE Robot Autom Lett 4(3):2296–2303
8. Riekert M, Klein A, Adrion F, Hoffmann C, Gallmann E (2020) automatically detecting
pig position and posture by 2D camera imaging and deep learning. Comput Electr Agric
174:105391. https://doi.org/10.1016/j.compag.2020.105391
9. Aichen W, Wen Z, Xinhua W (2019) A review on weed detection using ground-based machine
vision and image processing techniques. Comput Electr Agric 158:226–240 ISSN 0168–1699
10. Andreas K, Peta-Boldú FX (2018) Deep learning in agriculture: a survey, Comput Electr Agric
147:70–90 ISSN 0168–1699
11. Li S, Zhang Z, Du F, He Y (2020) A new automatic real-time crop row recognition based on
SoC-FPGA. IEEE Access 8:37440–37452
12. Imperoli M, Potena C, Nardi D, Grisetti G, Pretto A (2018) An effective multi-cue positioning
system for agricultural robotics. IEEE Robot Autom Lett 3(4):3685–3692
13. Vidović I, Cupec R, Hocenski Ž (2016) Crop row detection by global energy minimization.
Pattern Recogn 55:68–86 ISSN 0031–3203
14. Latif R, Saddik A, Elouardi A (2019) Evaluation of agricultural precision algorithms on
UAV images. In 2019 International Conference of Computer Science and Renewable Energies
(ICCSRE), Agadir, Morocco, pp 1-4
15. Horstrand P, Guerra R, Rodríguez A, Díaz M, López S, López JF (2019) A UAV platform
based on a hyperspectral sensor for image capturing and on-board processing. IEEE Access
7:66919–66938
16. Coakley C, Munro-Stasiuk M, Tyner JA, Kimsroy S, Chhay C, Rice S (2019) Extracting khmer
rouge irrigation networks from pre-landsat 4 satellite imagery using vegetation indices. Remote
Sens 11:2397
17. Ma D, Carpenter N, Amatya S, Maki H, Wang L, Zhang L, Neeno S, Tuinstra MR, Jin
J (2019) Removal of greenhouse microclimate heterogeneity with conveyor system for
indoor phenotyping. Comput Electr Agric 166:104979. https://doi.org/10.1016/j.compag.2019.
104979
18. Yang C (2020) Remote sensing and precision agriculture technologies for crop disease detection
and management with a practical application example. Engineering 6(5):528–532
19. Radoglou-Grammatikis P, Sarigiannidis P, Lagkas T, Moscholios I (2020) A compilation of
UAV applications for precision agriculture. Comput Netw 172:107148. https://doi.org/10.1016/
j.comnet.2020.107148
20. Tian H, Wang T, Liu Y, Qiao X, Li Y (2020) Computer vision technology in agricultural
automation —a review. Inf Process Agric 7(1):1–19
21. Roy R (2018) «ODROID-XU4 Manual». ODROID Magazine. https://magazine.odroid.com/
odroid-xu4
22. https://www.sensefly.com/education/datasets/?dataset=5632&industries%5B%5D=2
23. https://www.ipb.uni-bonn.de/data/uav-sugarbeets-2015-16/
24. Xue J, Titterington DM (2011) t -Tests, F -tests and otsu’s methods for image thresholding.
IEEE Trans Image Process 20(8):2392–2396
25. Blaschke T (2010) Object based image analysis for remote sensing. ISPRS J Photogrammetry
Remote Sens 65(1):2–16
26. Saddik A, Latif R, Elhoseny M, El Ouardi A (2021) Real-time evaluation of different indexes
in precision agriculture using a heterogeneous embedded system. Sustain Comput Inf Syst
30:100506. https://doi.org/10.1016/j.suscom.2020.100506
Design and Implementation of an Open
Source and Low-Cost Nanosatellite
Platform
Abstract In recent years, nanosatellites and CubeSats market have grown consider-
ably, making it possible to carry out, at low cost and in a short time, a wide range of
spacecraft missions (industrial or scientific) in Low Earth Orbit (LEO). In this paper,
we propose the use of the open-source movement to space technology by designing a
low-cost hardware and software platform for LEO nanosatellite. The hardware design
is based upon open source and flexible Arduino technology. The idea is to propose a
clear, simple, ease of programming, rapid prototyping, and publicly available hard-
ware platform. However, the standard Arduino API is restricted to the capabilities
found on a core set of library function, lacking support for multithreaded programs or
specification of real time requirements. So, to take maximum advantage of the hard-
ware platform performance, with a more memory-efficient programming, we have
implemented a FreeRTOS based software platform. The development environment
used was Atmel Studio and the AVR toolchain, which made it possible to directly
program the Arduino AVR Atmega microcontroller in pure C language.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 421
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_39
422 A. Hanafi et al.
1 Introduction
Triggered by the success of the CubeSat standard, the past ten years have seen the
nanosatellite segment grow by a factor of 10x, from as few as 20 satellites in 2011
to nearly 200 in 2019, and it is estimated that 1800 to 2400 nano/microsatellites
will need to be launched over the next 5 years [1]. Most of the recent CubeSats
were launched by commercial operator. However, the academic world takes a largest
share among CubeSat developer [2]. From this, it is evident that academia has a
strong impact on the future of the space industry by enabling the CubeSat program
not only to be an ideal tool for teaching hands-on space technology, but also a great
chance to open access to space technology to a much wider audience, including
students from emerging and developing countries.
The work presented in this paper is the start of an open-source initiative for
CubeSat technologies with the intention to encourage and support Moroccan graduate
students to design, build, test, and operate satellites within their academic curriculum.
The motivation of this initiative comes from the at times frustrating experience of
having built the Masat1 University 1U CubeSat in an academic environment. Our
perception was that collaboration within the CubeSat community is at a minimum,
with the most projects that show design and test results but omit all specific details.
As a result, any new university team deciding to start a CubeSat project is facing the
challenge of starting their development from scratch or buying expensive CubeSat
products on the market.
In the following, we first explain in Sect. 2 the benefits of an “open-source design”.
Then, Sect. 3 describes the adopted modular architecture for the hardware platform
while Sect. 4 present the main features of the detailed design and implementation of
the real-time software platform. Finally, Sects. 5 and 6 outline the main results of
the software testing and conclusions with future work to improve and complete the
proposed open-source architecture.
• Improve reliability with freely available source code/design for inspection. Design
bugs and sources of previous CubeSat mission’s failure could be quickly detected,
and users can even suggest solutions to identified problems.
• Customize the open-source solutions to the specifics of a CubeSat mission since
it is easy to modify the shared software and hardware design files as desired.
• potentially reduce costs and design times, as well as the human and technological
resources required to successfully carry out a CubeSat mission.
Our initiative aims to provide an open-source reference hardware and software
platform for generic CubeSat mission. Partially based on the guidance from the
NASA System Engineering Manual [5], a top-down design approach was adopted to
translate mission objectives and requirements into subsystems, hardware compo-
nents, interfaces, and software tasks. Figure 1 summarizes the platform design
approach which is divided into four interdependent, highly iterative, and recursive
processes, resulting in a validated set of requirements and a validated design solution
that satisfies a set of stakeholder expectations.
As a reference framework for our work, we partially used the ECSS (European Coop-
eration on Space Standardization) standards, which freely provide [6] the necessary
reference documents and tools for the management of the design and its system
engineering. For example, in the analysis phase of the project, we applied the project
planning and implementation standard ECSS-M-ST-10C which encompasses the
424 A. Hanafi et al.
entire process to be carried out to plan and execute a space project in a coordi-
nated, efficient, and structured manner [7]. This process makes it possible to identify
the Project breakdown structures which constitute a common and unique repository
making it possible to identify the tasks and responsibilities of each actor.
Thus, the product tree, shown in Fig. 2, conceptually decompose a typical CubeSat
space segment into platform and payload, with the platform (subject of our work)
composed of the traditional subsystems, such as on-board computer, power system,
communication system and attitude control system.
The initial purpose of the concept for the Arduino open-source hardware
was to build a simple and low-cost microcontroller board that students could
easily use for projects. The great advantage of the Arduino platform is that the
hardware and software resources of the different boards and shields are completely
open source [8].
However, the use of a Commercial off-the-shelf (COST) open-source module
like Arduino in the space applications, is a challenge. Indeed, even if some CubeSat
missions have already successfully integrated Arduino modules in their design [9,
10], they do not share the open-source spirit of our initiative and indicate the
integration and implementation procedures of these modules.
Considering constraints related to the CubeSat dimensions, weight, and energy
consumption, especially for 1U formats, the main component of the platform
subsystem design will be the Arduino Nano module. The Arduino Nano is a board
of dimensions 45 × 18 mm2 , noticeably light (5 g) and easily weldable on electronic
Design and Implementation of an Open Source and Low-Cost Nanosatellite Platform 425
In this sub-section, we will discuss the adopted reference design solutions for
the nanosatellite hardware platform, more specifically the architecture and the
communication bus.
As shown in Fig. 3, we have opted for a distributed architecture where each
subsystem is equipped with a local microcontroller that will manage the subsystem
specific tasks as well as communication with the rest of the platform. This choice
offers the following advantages:
• The on-board computer is no longer a singular point of failure since its failure will
not necessarily lead to the failure of the entire system and/or the loss of contact
with the space segment.
• Independent design and testability of the subsystems which reinforces the open-
source spirit of the project. Indeed, the hardware interfacing of the subsystems
being standard, they can be developed independently by different work teams.
Regarding the choice of the communication bus, we had to choose between I2C
(Inter Integrated Circuit) and CAN (Controller Area Network). For the selection
criteria, energy consumption was favoured over robustness and reliability, as shown
in Table 1:
The choice of the I2C bus has the advantage of low consumption, ease of imple-
mentation and an open-source license. But it is a low-level protocol that must
be improved by implementing a high-level connection-oriented protocol based on
addressing and error checking.
426 A. Hanafi et al.
The onboard flight software of our nanosatellite platform presents the characteristics
of a real-time system, as it must manage various external events and perform tasks
with time constraints. The use of a real-time operating system such as FreeRTOS
makes it possible to meet this requirement by giving the possibility of having a
428 A. Hanafi et al.
multi-tasking, modular and more reliable code to ensure the smooth operation of the
platform.
FreeRTOS is a small footprint real-time kernel for microcontroller. It is released
under the free GNU General Public License (GPL) and it offers different ports,
notably for ATmega323, but not for ATmega328. We then use the open-source repos-
itory of files “AVR ATmega port of freeRTOS” which can be flexibly configured uses
timers available on AVR ATmega devices to work with Arduino device [11].
For the port of the most recent version 10.4.1 of freeRTOS:
• The “freeRTOSBoardDefs.h” file contains most of the necessary variables to
change hardware platform. We define the variable allowing the use of the 328p
Timer0:
• The configuration file “FreeRTOSConfig.h” tailors the RTOS kernel to the devel-
oped software platform. We especially distinguish the clock frequency of the
microcontroller which is indicated by the F_CPU constant which was defined by
the Atmel Studio environment:
• There are some relevant and often used libraries added to the basic freeRTOS
capabilities, especially the “lib_io” that contains the tools to use the I2C bus.
The file “i2cMultiMaster.c” contains integrated interrupt driven master and slave
routines.
A connection-oriented protocol was designed on top of the I2C bus to allow commu-
nication between the different subsystems of the nanosatellite platform. This protocol
uses a transmit method separates between three types of packets as shown in Fig. 6:
Each type of frame begins with the destination sub-system address followed by
an identifier field which indicates the type of the frame:
• Write Data Frame: Used when transmitting data. The 2 most significant bits of
the identifier are set to “00” and the remaining 5 bits indicate the device module
number.
430 A. Hanafi et al.
• Read Data Frame: Used when requesting data from the receiver. The 2 most
significant bits of the identifier are set to “01” and the remaining 5 bits indicate
the device module number.
• Request Frame: Used when requesting the receiver for doing some commands.
The 2 most significant bits of the identifier are set to “10” and the remaining 5
bits indicate the command to be executed followed if necessary, by parameters.
This approach allows software designers to choose the commands and data formats
relevant to the tasks of each subsystem.
A real-time freeRTOS based application was created in the Atmel Studio develop-
ment environment to test the port of the operating system, the different functions of
our on-board computer library, and communication via the I2C bus.
The application has been implemented and tested in the Arduino nano module of
the on-board computer of the hardware platform shown in Fig. 7:
The total flash footprint came up to approximately 18 Kbyte which includes Free
RTOS, application and library code. A simple debug console interface provides a set
of text commands to test the implementation as shown in Fig. 8:
Design and Implementation of an Open Source and Low-Cost Nanosatellite Platform 431
This work is the beginning of an ongoing project to use open design to make space
technology accessible. Our team was thus able to develop in a noticeably short time
(3 months), the basis of an open-source hardware and software architecture that could
be used as a platform for CubeSats projects.
This was possible by using Commercial off-the-shelf products (Arduino modules)
and by adapting and enriching existing open-source software libraries. The encour-
aging results of this work have allowed us to set the following steps as future
work:
• Design a MPPT and Arduino Nano based smart electrical power system for our
nanosatellite platform.
• Design of the on-boards computer and power supply PCB (Printed Circuit Board)
by respecting the mechanical interfacing imposed by the PC104 standard and the
dimensional constraints (10 × 10 cm2 ) imposed by the CubeSat standard.
432 A. Hanafi et al.
• Improve the debugging shell interface used in this work and implement CRC
check in the communication protocol over I2C Bus, to detect accidental changes
to raw data during the transfer.
• Provide a complete repository containing the history of our project source files.
It will be hosted in a single Git repository like GitHub.
References
1. Williams C, DelPozzo S (2020) Nano microsatellite market forecast. 10th Edition, SpaceWorks
Annual Nano/Microsatellite Market Assessment
2. Erik Kulu, Nanosatellite Database. https://www.nanosats.eu/index.html#database. Accessed 5
Oct 2020
3. Open-source initiative. https://opensource.org/. Accessed 10 Oct 2020
4. Open-source hardware association. https://www.oshwa.org/. Accessed 10 Oct 2020
5. National Aeronautics and Space Administration NASA Headquarters (2007) “Systems
Engineering Handbook”. NASA/SP-2007–6105 Rev1
6. European Cooperation for Space Standardization. https://www.ecss.nl/. Accessed 10 Oct 2020
7. ECSS-M-ST-10C, (2009) Project Planning and Implementation. ESA-ESTEC
8. Arduino Open-source initiative. https://arduino.cc/. Accessed 10 Oct 2020
9. Geeroms D, Bertho S, De Roeve M, Lempens R, Ordies M, Prooth J (2015) ARDUSAT, an
Arduino-based cubesat providing students with the opportunity to create their own satellite
experiment and collect real-world space data. In 22nd ESA Symposium on European Rocket
and Balloon Programs and Related Research, vol 730, p 643
10. Roman-Gonzalez A, Quiroz-Olivares AE, Vargas-Cuentas NI (2020) Advances in the UCHSat-
1 nanosatellite: design and simulation. Adv Astronaut Sci Technol 3:65–74. https://doi.org/10.
1007/s42423-020-00054-1
11. FreeRTOS & libraries for AVR ATmega. https://github.com/feilipu/avrfreertos. Accessed 20
Oct 2020
Design and Realization of Fire Safety
System for Controlling and Monitoring
a Siren Using Arduino Uno
Abstract In this work, the design and realization of fire safety system for controlling
and monitoring a siren using Arduino Uno are proposed for decreases the number
of risks in the evacuation process to protect the medical team and the patients living
in the Moroccan University Hospital Centers. To solve the problem of controlling
and monitoring the siren in the existence of fire, we have proposed a new system
to control the siren during a fire and transmit its status to the Control and Signaling
Equipment, this new system will be placed between the ATCAV control system and
the siren. We used the Arduino UNO, voltage sensor, H-bridge, voltage regulator,
Bipolar Transistor NPN TIP120, Optocoupler, LCD and siren in the hardware tools.
And also we used the Arduino IDE and EasyEDA in the software tools.
Keywords Fire safety system · Arduino Uno · H-bridge · Siren · Sensor voltage
1 Introduction
As we all know the fire appears since the existence of human beings on earth, it
is among the materials necessary to nourish the needs and daily requirements of
humanity. Moreover, fire also has a bad history with humanity because of the fires
caused lately in the different parts of our world. For that, it has always been a threat
to human life.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 433
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_40
434 A. Morchid et al.
The first intervention of humanity to control and extinguish the fire is being by
the use of water, but this solution is less effective in some cases. For this reason the
humanity looked for other solutions that depend on the technological development
to detect and control the risk of fire automatically.
The risk of fire can be very significant and cause loss of life. The only solution
to relieve these losses is to respond quickly to this emergency. In such situations,
the early detection of fire and the provision of a safety alert will protect loss of life
and property through the use of a fire alarm system that can be monitored locally or
remotely, as appropriate [1].
However, to be able to use the technology currently available, it will have to be
beneficial to the lives of others, which will be a great contribution to the community
[2, 3]. Although there are several well-developed systems available for use in practical
scenarios [4, 5] and [6], but also these systems are not cost-effective, unreliable in
automatic response, and not easy to implement. Today, in the market, there are
different types of fire alarm systems that are very complex in design, which requires
preventive maintenance to ensure its operation. In the meantime, if the system in
use requires maintenance to be performed each time, the cost of the use of it could
increase [1]. There are many researchers who have already worked on the fire safety
system to reduce the risks resulting from fire hazards [7, 8].
The fire safety system installed at the level of the Moroccan University Hospital
Centers (CHU) includes an ATCAV control system (Addressable Controllable Tech-
nical Alarm of the Vega range), this latter can only control the siren during the fire
but does not have the monitoring system for the siren.
In addition, if the fire exists and the siren fails at the same time at the university
hospital, in this situation the Control and Signaling Equipment (ECS) cannot trigger
and locate the fault, which increases the number of risks in the evacuation process
to protect the medical team and the patients living at the university hospital.
To solve this problem, we have proposed a new Fire Safety System to control
the siren during a fire and transmit its status (working, corrupted) to the Control and
Signaling Equipment (ECS), this new system will be placed between the ATCAV
control system and the siren.
We presented this paper as the following: in Sect. 1 an Introduction; in Sect. 2 a
Description of the proposed Fire Safety System; in Sect. 3 a Description of the tools
used; in Sect. 4 a Simulation and hardware realization of the proposed Fire Safety
system; and in Sect. 5 a conclusion.
In case of fire a signal is sent from the Fire Detector to ATCAV which will switch the
relay armature of the ATCAV to the power supply pin, consequently a command is
sent to the Fire Safety System which will switch on the siren. The principal functions
of the proposed Fire Safety System are illustrated in Fig. 1.
Design and Realization of Fire Safety System … 435
The ATCAV is connected to the siren by the NO pin of its relay and the common
(COM) pin is connected to a 24 V power supply. In case of fire, a signal sent by the
control unit allows the relay armature switched to the 24 V pin, which operates the
siren.
The ATCAV installed in Moroccan CHU can only control the siren during the fire,
it never has the monitoring part of the siren, that is to say that the contact between
pins M and J of the ATCAV will always remain closed as illustrated in Fig. 2.
On the part of monitoring the siren, our proposed Fire safety System is based
on the measurement of the voltage at the terminal of a resistor with the siren and
their polarity resistance in the middle of H-bridge, the purpose of using an H-bridge
for reverse polarity direction. In addition, the part of controlling is assured by an
Arduino Uno card that has been programmed to control the siren.
The siren is powered by 24 V, when the fire is declared by the Fire Detector block,
the normal polarity is applied. That means the positive terminal of the voltage source
in the middle of the H-bridge is connected with the positive terminal of the siren and
the negative terminal is connecting to the negative terminal, this polarity allows to
measure a voltage to determine the state of monitoring as shown in Fig. 3.
436 A. Morchid et al.
On the other hand, when the fire is not declared by the Fire Detector block, the
reverse polarity is applied. That is means the positive terminal is connected to the
negative terminal and the negative terminal is connected to the positive terminal of
the siren as shown in Fig. 4.
The hardware tools used in our proposed Fire Safety system are Arduino Uno, voltage
sensor, H-bridge, voltage regulator, Bipolar Transistor NPN TIP120, Optocoupler,
LCD and siren. On the other hand the software tools used are Arduino IDE and
EasyEDA.
The Arduino system is a tool for making small computers that can control more things
in the hardware world than your desktop computer [9]. It is an open source platform
of programs, electronics that are based on a microcontroller card and software, a true
integrated development environment, to write, compile and transfer the program to
the card.
Among all the cards of the Arduino family, we chose the UNO version because
of its characteristics and it is from this base that we developed the whole project.
Design and Realization of Fire Safety System … 437
The Uno model from Arduino is an electronic card whose heart is an ATMEL
microcontroller of the ATMega328 reference. The ATMega328 microcontroller is an
8-bit microcontroller from the AVR family, which can be programmed in C language.
The Arduino Uno card as shown in the Fig. 5 above is a microcontroller card based
on the ATmega328. It has: 14 digital input/output pins (6 of which can be used as
PWM (pulse width modulated) outputs), 6 analogue inputs (which can also be used
as digital input/output pins), a 16 Mhz quartz crystal, a USB connection, a power
jack connector, an ICSP connector (“in-circuit” programming) and a reset button.
Figure 6 shows the voltage sensor V < 25 V. This module is based on a resistive
design divider principle, can reduce the voltage of the input terminal connected five
times, Arduino analog input voltage up to 5 V. The main characteristics of it are: an
input voltage range (DC 0–25 V), a voltage detection range (DC 0.02445–25 V), an
3.3 H-Bridge
The H-bridge allows a current to flow through a load in both directions as illustrated
in Fig. 8. It is used to run a DC motor in both directions and stop it (motor controlled),
to create an AC voltage (inverter) and to create an adjustable DC voltage (chopper)
[12].
A H-bridge is composed of 2 parts, a power part composed of 4 SW-controlled
switches (relay, bipolar transistor or MOS…) and a control part which controls the
opening (O) or closing (F) of the switches. This control is most often done like PWM
(pulse width modulation).
The objective of using an H-bridge based on NPN TIP120 bipolar transistors
in our project to reverse the polarity direction from the idea of reversing the two
directions of a load (Siren).
An LCD display is capable of displaying all the usual alphanumeric characters and
some additional Symbols as shown in Fig. 9. For some displays it is even possible
to create your own characters [13].
The Bipolar Transistor TIP120 of NPN type is shown in Fig. 10. It is used to switch
high current loads (up to 5A), as a medium power switch with high amplification is
required [14]. And also it is used in speed control of motors and inverter and other
rectifier circuits.
The voltage regulator LM317 is shown in Fig. 11. The LM317 is a linear voltage
regulator. This family of integrated circuits allows, for any input voltage, to provide
a fixed output voltage from which current can be drawn. It is a 3-leg serial regulator
capable of providing 1.5 A on an output voltage ranging from 1.25 to 37 V. It is
exceptionally easy to use and requires only 2 external resistors to fix its output
voltage. In addition, the power supply rejection and insensitivity to load impacts
make it an advantageous regulator [15].
The AVS 2000 sound diffuser is a sound evacuation device for fire. It can be
integrated into the Alarm Equipment (EA) provided for in the NFS 61 936 standard
and can be controlled by an Alarm Management Unit (AGU). The Siren AVS2000-
SIP is shown in Fig. 12.
It is designed around piezoelectric technology, AVS 2000 guarantees very low
power consumption. Its circular design and its side vents allow it to broadcast sound
in several planes, ensuring perfect hearing of the alarm throughout the volume [16].
4.1 Simulation
Figure 13 shows the schematic of our proposed Fire Safety System in EasyEDA
software. And Fig. 14 shows the schematic of the power Supplies system, Voltage
Sensor and input pins in same software.
We have used the LM317T regulator in schematic to adapt the input power supply
voltage that equal 24 V to an adapted voltage at 9 V which used to power the Arduino
Uno Card. And also the voltage sensor used to measure the voltage at the resistor
terminal.
In the power part, the siren and the polarity resistor located in the middle of the
H-bridge as illustrated in Fig. 15.
In the control part, the Arduino Uno Card with the connection connectors and
all the components as well as the output pin and also the part who declares to the
proposed Fire Safety System if there is a fire and the status of the siren, are illustrated
in Fig. 16.
Design and Realization of Fire Safety System … 441
Fig. 13 The schematic of the proposed fire safety system in EasyEDA software
We have used a 24 V DC power supply for the siren which powered during a fire.
When there is no fire the voltage sensor will measure a zero voltage which means a
disconnection for the proposed Fare Safety system.
The practical realization of the voltage regulator is illustrated in Fig. 17, we have
used a multimeter to show the actual output value by using two resistors (R1 = 220
and R2 = 1.5 ) that regulate the output value at 9.5 V and also two capacitors to
suppress parasites (noise) in the output.
If the fire existing and the siren on we will have that the push button is pressed,
the normal polarity on the H-bridge is +24 V, Either the polarity resistance exists
or not, the measured voltage is 8.64 V when R is ON and 7.42 when R is OFF. line
ok = working and green LED on, and no change in the contact between the M and j
pins of the ATCAV as there is no fault as illustrated in Fig. 18. If else the fire existing
and the siren off we will have that the push button is pressed, the normal polarity on
the H-bridge is +24 V, the line OFF = default, the red LED lights up, indicating that
a disconnection has occurred, open the contact between pins M and J of the ATCAV,
then the ECS will declare the existence of a fault accompanied by the address of
the ATCAV, and the value of the voltage 0.02 V with the type of fault at the siren is
displayed on the LCD screen as illustrated in Fig. 19.
If the fire not existing and the siren on we will have that the push button is not
pressed, the application of reverse polarity on the H bridge is −24 V, Either the
polarity resistance exists or not, the measured voltage is 7.20 V when R is ON and
5.59 V when R is OFF, the line ok = working and green LED lit and no change in
the contact between pins M and j of the ATCAV since there is no disconnection as
illustrated in Fig. 20. If else the fire not existing and the siren off we will have that
the push button is not pressed, the application of reverse polarity on the H bridge is
−24 V, the red LED lights up to indicate disconnection, the line OFF = fault and the
measured voltage is 0.04V, open the contact between pins M and J of the ATCAV
and then the ECS will declare the existence of a fault accompanied by the address of
the ATCAV as illustrated in Fig. 21.
5 Conclusion
The proposed Fire Safety System have been successfully designed and realized to
controlling and monitoring a Siren using Arduino Uno. Simulation, analysis and
practical realization are all included in this study. It significantly solve the problem
of controlling and monitoring the siren (transmit its status) in the existence of fire. The
proposed Fire Safety System implemented in Arduino Uno by using a voltage sensor,
an H-bridge, a voltage regulator, a Bipolar Transistor NPN TIP120, an Optocoupler,
LCD and a siren. The new system simulated in Arduino IDE and EasyEDA software
tools.
References
1. Suresh S, Yuthika S, Vardhini GA (2016) Home based fire monitoring and warning system. In:
2016 international conference on ICT in business industry & government (ICTBIG), Indore,
pp 1–6. https://doi.org/10.1109/ICTBIG.2016.7892664
2. Zeebaree SR, Yasin HM (2014) Arduino based remote controlling for home: power saving,
security and protection. Int J Sci Eng Res 5(8):266–272
3. Piyare R, Tazil M (2011) Bluetooth based home automation system using cell phone. In:
15th international symposium on consumer electronics (ISCE), Singapore. IEEE, pp 192–195.
https://doi.org/10.1109/ISCE.2011.5973811
4. Jinghong L, Xiaohui Z, Lu W (2012) The design and implementation of fire smoke detection
system based on FPGA. In: 24th Chinese control and decision conference (CCDC), Taiyuan,
pp 3919–3922. https://doi.org/10.1109/CCDC.2012.6244626
5. Celik T, Demirel H, Ozkaramanli H, Uyguroglu M (2006) Fire detection in video sequences
using statistical color model. In: IEEE international conference on acoustics speech and signal
processing proceedings, Toulouse, p II. https://doi.org/10.1109/ICASSP.2006.1660317
6. Angayarkkani N, Radhakrishnan N (2009) Efficient forest fire detection system: a spatial data
mining and image processing based approach. Int J Comput Sci Netw Secur 9(3):100–107
7. Song WS, Hong SH (2007) A reference model of fire detection and monitoring system using
BACnet. J Build Environ 42(2):1000–1010. https://doi.org/10.1016/j.buildenv.2005.10.028
8. Sharma S, Chand K, Sharma D, Guha P (2018) Development of an early detection system for
fire using wireless sensor networks and Arduino. In: International conference on sustainable
energy, electronics, and computing systems (SEEMS), Greater Noida, India, pp 1–5. https://
doi.org/10.1109/SEEMS.2018.8687333
9. https://components101.com/microcontrollers/arduino-uno
10. https://components101.com/sensors/voltage-sensor-module
11. https://learnabout-electronics.org/Downloads/PC817%20optocoupler.pdf
12. https://www.build-electronic-circuits.com/h-bridge/
Design and Realization of Fire Safety System … 445
13. https://components101.com/16x2-lcd-pinout-datasheet
14. https://components101.com/transistors/tip120-pinout-datasheet-equivalent
15. https://components101.com/lm317-pinout-equivalent-datasheet
16. https://studylibfr.com/doc/901746/fiche-tech-avs2000-mise-en-page-1.qxd
Water-Purifying Distiller with a Cooling
System Controlled by a Photovoltaic
Panel
Abstract Globally, the demand for good quality drinking water is growing. Indeed,
the population is increasing rapidly and the water needs of industry and agriculture are
higher and higher. This paper presents the theoretical study and practical realization
of a system which consist to desalinate and purify sea water, equipped with a cooling
system powered with an automatic photovoltaic panel sun tracker. It’s a cheaper
design, simple and robust, adapted to the needs and specificities of our vast country.
This equipment has been tested during the months of December and October 2020.
The daily production recorded in distilled water varies from 600 to 645 ml under a
daily radiation received on the plan and vary within a range of 5.0 to 5.3 kWh/m2
under an ambient temperature which varies from 21.6 to 33.5 °C.
1 Introduction
By 2022, water consumption [1] is expected to increase by 40% and half the planet
will lack clean water [2]. Today, no less than 900 million people do not have access
to drinking water. Each year, eight million people, including 1.8 million children, die
from pathologies linked to the consumption of unsanitary water, diarrhea, cholera,
typhoid, malaria, etc. [3].
For these facts, we have been encouraged to think about a solution: That is the
sea water desalination system [4]. To explain the technique, seawater distillation is
boiling followed by liquefaction. This process is therefore based on a phase change:
The seawater solution is introduced into the heated flask. When it boils, it turns into
vapor and rises in the air column. The water molecules vaporize, thus separating
from the chloride ions Cl− and sodium Na+ [5]. The water then flows through a
cooling tube to return to a liquid state and falls as droplets into the beaker, forming
a pure water distillate [6]. Thus, through this process, the water is separated from
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 447
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_41
448 K. Mounich et al.
the salt and other microbes that remained in the initial solution. Ahmed Khedim [7]
studied in this context a seawater distiller with multiple stages and heat recovery. It
is therefore a question of finding an operating mode, allowing the distiller to obtain
the best efficiency, when thermal energy is recovered several times. This method of
recovery and use of several times the same amount of solar energy captured has for
aim to significantly reduce energy consumption. Experience shows that the system
provides a good return. The idea developed by the authors is to achieve using simple
technical means based on solar energy and with a minimum cost, a modest supply of
drinking water. This method used, involves recovering and using the same amount of
thermal energy several times. It has for aim to lower energy consumption so that solar
energy becomes an alternative and competitive source among all water desalination
technology techniques.
Ali A. Badran [8] presented a purely experimental study on a solar still connected
to a solar collector to study the effect of increases in settings on the still under local
conditions. The results show that productivity increases significantly compared to
that of the distiller alone.
Rajesh Tripathi and G.N. Tiwari presented an experiment to determine the convec-
tive heat transfer [9] coefficient for the active solar distillation system. They found
that the coefficient of convective heat transfer between the water and the air surface.
The interior of the condensation cover depends significantly on the water depth.
Also, they observed that the yield obtained is important over the hours of sunshine
compared to the daily production for water depths in the solar still (0.10 m and
0.15 m), due to the storage effect. The active still is more efficient than the passive
still but their cost is very high and their use is more complex. In the experiment
we presented, the heat source is provided by a butane cylinder, however, it must
be replaced by a solar heat source to minimize transformation cost and avoid the
complexity of the process.
2 Experimental Part
2.2 Experiences
The experience is spread over a period of 2 months (September, October). The water
of the process comes from sea water. The stainless-steel container is filled with 700 ml
of water, 2/3 of the volume, the salinity of the water is 28.7 g [10]. The increase in
temperature due to the thermal effect which heat the saline water which evaporates.
This evaporation capacity increases as the temperature increases to boiling. The
state of water changes from liquid to gas. The water vapor in the hot air is cooled
through the passage to the copper solenoid; the vapor condenses inside the copper
tube (Fig. 1). There is a formation of water droplets which flow through the spiral
which leads it to an activated carbon filter [11].
The experimental study will focus on the measurement of the following param-
eters: Global solar radiation on an oblique plane is measured using a pyranometer.
The different temperatures are measured using type K thermocouples (Chromel -
Alumel). These climatic parameters of the atmosphere are taken at the end of each
hour by an instrument called Testo 925. The water salinity measurement is done
using the HI98319 Salinity Tester, the water purity measurement is done using the
TDS meter (Total Dissolved Solid) (Table 1).
450 K. Mounich et al.
The electronic part is the key milestone of this project success. In fact, the electric
power is ensured by the photovoltaic panel to all electronic items. The type of photo-
voltaic panel is monocrystalline 5 V 12 W. Below are both electronic schema and
real prototype picture (Fig. 2).
We have coupled the two photovoltaic panels in parallel. A parallel mounting of two
cells increases the generator output current thus created. In a group of identical cells
Water-Purifying Distiller with a Cooling System Controlled ... 451
connected in parallel, the cells are subjected to the same voltage and the characteristic
resulting from the grouping is obtained by adding the currents, Fig. 3. Equation (1)
in turn summarizes the electrical characteristics of a parallel association of NP cells.
Icc = N ∗ Icc (1)
Figure 4 shows the equivalent diagram of a photovoltaic cell under lighting. It corre-
sponds to a current generator mounted in parallel with a diode. Two parasitic resis-
tances are introduced in this diagram. These resistances have a certain influence
452 K. Mounich et al.
on the characteristic I = f (V) of the cell: The series resistance (Rs) is the internal
resistance of the cell; it mainly depends on the resistance of the semiconductor used.
The shunt resistance (Rsh) is due to a leakage current at the junction; it depends
on how it was mounted (Fig. 5).
The mathematical model for the current-voltage characteristic of a PV cell is given
by the formula (2) [13].
q* Vpv + Ipv * Rs Ipv *Rs + Vpv
Ipv = Iph − Isat exp + −1 − (2)
a*K*T*N Rsh
Where:
Iph : light generated current of PV module in Amperes
Isat : reverses saturation current in Amperes
Rs : series resistance in Ohms
Rsh : parallel resistance in Ohms a Diode ideality factor
K: Boltzmann constant (1.38 × 10−23 J/K)
q: electron charge (1.6 × 10−19 C)
N: number of series connected cells in a module
T: module temperature in Kelvin
Ipv : output current of PV module in Amperes
V: output voltage of PV module in Volts
was detected at 10:30 a.m. and the total amount of distillate collected until 12 p.m.
is equal to 645 ml (Figs. 6 and 7).
200
100
0
10:40 10:50 11:00 11:10 11:20 11:30 11:40 11:50 12:00
Among the results we did the physico-chemical test of the fresh water coming
from the filter with the TDS device and we had the following results (Fig. 8):
4 Conclusion
The role of this project is to give a clear vision of the importance of using the
renewable energies (solar energy) as a key milestone to solve the problem of the
scarcity of drinking water in certain critical areas, and in the rest of the world in
general. Indeed, the use of solar energy offers a gigantic pure source and reliable
with competitive cost.
The problem of producing fresh water by solar means is important to two
Viewpoints: need to save energy and concern for new water resources.
This paper presents the experimental evaluation of the main parameters influ-
encing the operating characteristics of a solar distiller with photovoltaic panels.
The results obtained showed that the production, the overall efficiency of the Solar
still increases with solar irradiation provided by the servo motor and photo resistors
(Sun follower) and the efficiency of the copper tube cooling system.
Following the various experiments carried out on our solar distiller with photo-
voltaic panels, we found that to improve its performance and make it efficient, it will
take:
– Heat gain can be provided by preheating the feed water from a solar thermal
source.
Water-Purifying Distiller with a Cooling System Controlled ... 455
References
1 Introduction
The researchers found that the security architecture utilized in 4G did not meet the
needs of 5G. Although, some researchers have proposed a sort of inflexible unified
network security architecture, the security architecture must always be adapted
to the needs of the 5G eMBB (enhanced Mobile Broadband), mMTC (massive
Machine Type Communications), and URLLC (Ultra-Reliable and Low Latency
Communications) application scenarios. They also proposed a physical layer authen-
tication scheme dependent on the hash technique. Despite the fact that this tech-
nique is light, the application scenarios are limited. Whereas, others are conducting
cross-authentication studies.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 457
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_42
458 H. Amgoune and T. Mazri
Nowadays, the security between user equipment (UE), base stations, and core
networks is highly needed, furthermore, between the UE and the base station, inter-
layer security measures are appropriate, whereas, the physical layer security measures
are perfect between the base station and the main network. In the case of UEs (D2D
(Device-to-device) pairs or UE), the inter-layer security measures and the traditional
security measures are adopted to the circumstances. Moreover, physical layer security
measures work best in low-speed cases, thus, in high-speed cases, they have to face
inter-layer security measures. Many questions deserve to be studied and explored in
the field of security in 5G. In particular, can we use cryptographic algorithms on all
devices in5G services (eMBB, mMTC, and URLLC)? It should be noted that many
of these devices cannot have additional circuitry, power consumption and code space
required to process data for cryptographic methods due to the limited resources of
some 5G devices. The rest of this paper is organized as a following. In Sect. 2, we
cover security at the application layer, in Sect. 3, we present security methods at the
physical layer. In Sect. 4, we present the discussion. The conclusion is formulated in
Sect. 5.
Ensuring security between IoT devices and MEC (Multi-Access Edge Computing)
servers is a challenge because we are talking about both physical and superior secu-
rity. Security mechanisms must be applied to protect the offload process from espi-
onage attacks when IoT devices delegate computational tasks to MEC servers. Partic-
ularly complicated, especially for IoT devices that have limited resources. To do this,
security mechanisms must be integrated into the MEC architecture. The cooperative
calculation and transition procedure consists of two stages:
• Safe execution of account tasks.
• Secure transfer of account results.
Appear IoT applications such as wearables i.e. electronic devices which have the
possibility to store and process user data. These devices are integrated into the clothes
or accessories of the person wearing them. These devices store and manage sensitive
information about humans and intelligent healthcare. The security drawbacks of these
wearables are numerous, as an example, the leakage of personal information is cited,
which will impact the privacy of the person.
Previously, IoT security was enhanced by data encryption at the upper layer of
the communication protocol stack. However, generating, distributing and managing
secret keys can be complicated and difficult to implement large-scale IoT networks
especially since there are many complex nodes like sensors.
According to the researchers [1], the 5G network should support differentiated
scenarios and various wireless services. Indeed, different types of services have very
different security requirements. For example, online payment requires a much higher
level of security than usual. However, encryption-based methods can only provide
Security at the Application Layer and the Physical Layer in 5G 459
“binary” levels of security. In other words, the information is protected if the secret
key is exchanged securely.
According to the same researchers, IoT applications presented by machine type
communications (MTC) must be supported by 5G, where huge MTC devices run out
of energy, storage and computing capacity, and the complex encryption/decryption
algorithms or protocols cannot be applied [1].
For the reasons below these researchers have proposed to use physical layer secu-
rity solutions (PLS) which take full advantage of the characteristics of wireless
channels to degrade the qualities of the signal received by malicious users and to
realize transmission. Secure without keys via signal design and signal processing
techniques. PLS avoids the difficulties of distributing and managing secret keys
and offers flexible security levels thanks to the design of an adaptive transmission
protocol. In addition, PLS techniques correspond well to the characteristics of 5G
networks. The most promising solution to avoid security threats at the 5G network
level is the application of PLS [1].
The physical layer provides the best wireless network security [2]. There are many
security methods at the physical layer level, we mention:
In the article [3], the researchers proposed a more flexible security method that can
be adapted to the needs of the 5G application scenarios, namely eMBB, mMTC
and URLLC. They proposed to introduce a light authentication scheme between
layers below the proposed scheme. Channel state information and the cross-layer
authentication based on EAP/AKA protocol proposed by [3] (Fig. 1).
Extensible Authentication Protocol or EAP is a network communication protocol
embedding multiple methods of partner authentication, i.e., it is designed to provide
an authentication mechanism to authorize or prohibit the establishment of the
connection. EAP-AKA (Authentication and Key Agreement) is an EAP method for
customers of mobile telephone networks. This method allows mutual authentication
of the user equipment and the network. This protocol is mainly used in the following
settings:
• Authentication when attaching to a mobile phone network.
• Reuse of authentication elements from the mobile network when attaching to a
wireless network which belongs to the operator.
460 H. Amgoune and T. Mazri
A new waveform frame was proposed by Huawei called filtered OFDM (f-OFDM) to
enhance spectrum usage in eMBB and enabling flexible waveform for 5G. The main
task of f-OFDM is to divide the bandwidth into multiple sub-bands so that different
services can be hosted in different sub-bands.
The F-OFDM inherited all the benefits of the OFDM, such as high spectral effi-
ciency and MIMO adaptability, but it has also solved some of its flaws by improving
flexibility and spectral efficiency. It becomes a key technology to cut the 5G air
interface [4].
In [3], the authors proposed a new encryption scheme assisted by physical layer
security which is applied to the f-OFDM system (Fig. 2).
optical signals while ensuring high security. QKD is based on the principles of
quantum physics and is not based on mathematical or computational assumptions
[5].
This quantum cryptography makes it possible to distribute a secret encryption key
between two distant interlocutors, while ensuring the security of the transmission
thanks to the laws of quantum physics and information theory. This secret key can
be used afterwards in a symmetric encryption algorithm, to encrypt and decrypt
confidential data.
We use the quantum cryptography system to transmit a key, and not the message
itself for two essential reasons:
• The bits of information communicated by the mechanisms of quantum cryptog-
raphy can only be random. This is not suitable for a message, but perfectly suited
for a secret key, which must be random.
• Even if the mechanism of quantum cryptography guarantees that spying on
communication will always be detected, it is possible that the spy intercepts
bits of information before it is detected. This is unacceptable for a message,
but unimportant for a random key, which can be simply discarded in the event of
interception.
However, a smooth integration and wide adoption of QKD systems in telecom-
munications networks is an important step towards a practical and economical use of
quantum cryptography, which is still difficult due to technical and economic obstacles
[5]. Indeed, its physical implementation has been a problem for scientists. Several
462 H. Amgoune and T. Mazri
types of listening methods have been introduced, each with its own advantages and
limitations [6].
The future fifth generation (5G) wireless network is envisioned to be very heteroge-
neous and capable of supporting future applications requiring transparent coverage,
low latency and high speed. These requirements pose strict challenges in the design of
transfer authentication. In article [7], the researchers presented physical layer authen-
tication. They proposed to use transfer authentication with Kolmogorov-Smirnov (K-
S). To summarize, this work presents an authentication scheme based on the phys-
ical layer for SDN-compatible 5G HetNets. The proposed authentication scheme
consumes less computing and storage resources compared to the GLRT methods
according to the researchers [7]. Simulations have shown that the proposed K-S test
based on the scheme can provide reliable security performance for the purpose of
additional authentication [7].
On the other hand, researchers [8] proposed the Kolmogorov-Smirnov test for
user identification and an authorization procedure based on the typing dynamics
of users. In the method proposed by these researchers, the underlying time series
is preprocessed using an adaptive filter. They showed that the proposed method
performs much better than the basic classifier (1 NN DTW).
The researchers in the paper [9] proposed the digital radiofrequency (RF) tattoo
which is a new physical layer authentication technology, the carrier of which is
not traditional multimedia signals but digital signals. A blind watermark detection
scheme with Kolmogorov-Smirnov (KS) test is used.
Security-based power control and resource allocation technologies can adjust the
transmitter parameters to the instantaneous CIS, so that the SNR received from
legitimate users can be improved or maintained, while that of listening varies
randomly.
The perfect CSI of the legitimate channel [10] is not available on the transmitter
side due to a delay or an estimation error of the channel. To solve this problem,
researchers [10] used the knowledge contained in obsolete CSIs to decide whether
to transmit.
A channel amplitude quantification method that can use a single threshold value to
efficiently quantify the channel response is proposed in paper [11] in order to extract
a random manipulation sequence with good secrecy properties. More precisely, a
time division duplex (TDD) wireless system is considered to be independent iden-
tical distributed (iid) distribu Rayleigh fast discoloration channel (Rayleigh is a mode
of diffusion of waves, for example electromagnetic or sound, whose length wave is
much larger than the size of the diffusing particles. We speak of elastic scattering,
because this is done without variation of energy, in other words the wave keeps the
464 H. Amgoune and T. Mazri
same wavelength), where potential passive eavesdropping (Eves) can only estimate
their own channel and have no knowledge of the CSI between the legitimate commu-
nication parties. The transmitter (Alice) only knows the CSI of the legitimate user
(Bob). In particular, the security technique proposed by these researchers [11] takes
the bits of the transmitted data packets and manipulates them with a logical vector
which characterizes the randomness of the channel as a function of the estimated
CSI gain. The manipulation process is implemented at the bit level using an XOR
operation exactly before the modulation process. The same XOR operation is imple-
mented after the demodulation process on the detected bits to extract the hidden bits.
According to the researchers, the simulation results obtained show that the use of
such a mechanism can guarantee the confidentiality of the data.
Relay is one of the useful techniques for improving the security of the wireless phys-
ical layer. The existing literature shows that the use of a full duplex relay instead of a
conventional half-duplex relay improves the capacity of secrecy and the probability
of interruption of the secrecy.
In this article, the successive two-way relay (TPSR) was proposed in [12] to
improve the security of wireless transmission. We compare the achievable ergodic
secret rate, the probability of interception and the secret probability of interruption of
the proposed TPSR compared to half-duplex relay (HDR), full-duplex relay (FDR)
and full-duplex interference (FDJ) existing. The numerical results presented in this
paper reveal that the proposed TPSR achieves the same achievable ergodic confiden-
tiality rate as the FDR. The TPSR also offers the lowest probability of interception
and interruption of secrecy compared to the other schemes due to the efficiency of
the bandwidth in full duplex and the two mutually independent links assisted by the
relay pair R1 and R2.. The probabilities of interception of the TPSR and the FDR
are analyzed and verified simultaneously. The analysis shows that the probability
of interception of the TPSR is quadratic lower than that of the FDR. In short, with
the proposed TPSR protocol, secure wireless transmission can be achieved using
conventional half-duplex relays without the use of sophisticated interference and/or
interference cancellation techniques.
4 Discussion
5 Conclusion
Nowadays, researchers pay Attention to 5G New Radio (NR) Physical Layer Security
Problems. This due to the introduction of all IP technology, the growing number of
subscribers, the interconnection of most networks to the Internet, and the needs of
high confidentiality into certain applications. The security of today’s smart devices
has become vulnerable to malware and other malicious software attacks, as these
devices are used in sensitive processes, such as banking, location, or even online
health applications.
Thus, further improvements and strengthening of security standards for 5G mobile
communications are needed. In recent years, there has been an increase in research
interest in the potential of the physical radio channel properties for providing secure
communications.
466 H. Amgoune and T. Mazri
References
1. Sun L, Du Q (2017) Physical layer security with its applications in 5G networks: a review.
China Commun 14:14
2. Mazin A, Davaslioglu K, Gitlin RD (2017) Secure key management for 5G physical layer
security. In: 2017 IEEE 18th wireless and microwave technology conference (WAMICON),
Cocoa Beach, FL, USA, April 2017, pp 1–5. https://doi.org/10.1109/WAMICON.2017.793
0246
3. Pan F, Jiang Y, Wen H, Liao R, Xu A (2017) Physical layer security assisted 5G network security.
In: 2017 IEEE 86th vehicular technology conference (VTC-Fall), Toronto, ON, September
2017, pp 1–5. https://doi.org/10.1109/VTCFall.2017.8288343
4. Up in the air with 5G - Huawei Publications. https://www.huawei.com/en/about-huawei/pub
lications/communicate/80/up-in-the-air-with-5g. Accessed 25 May 2020
5. Aleksic S (2015) Towards fifth-generation (5G) optical transport networks. In: 2015 17th
international conference on transparent optical networks (ICTON), Budapest, July 2015, pp
1–4. https://doi.org/10.1109/ICTON.2015.7193532
6. Gupta S, Sau K, Pramanick J, Pyne S, Ahamed R, Biswas R (2017) Quantum computation
of perfect time-eavesdropping in position-based quantum cryptography: quantum computing
and eavesdropping over perfect key distribution. In: 2017 8th annual industrial automation and
electromechanical engineering conference (IEMECON), Bangkok, Thailand, August 2017, pp
162–167. https://doi.org/10.1109/IEMECON.2017.8079582
7. Ma T, Hu F, Ma M (2017) Fast and efficient physical layer authentication for 5G HetNet
handover. In: 2017 27th international telecommunication networks and applications conference
(ITNAC), Melbourne, VIC, November 2017, pp 1–3. https://doi.org/10.1109/ATNAC.2017.
8215379
8. Ceffer A, Levendovszky J (2016) Kolmogorov-Smirnov test for keystroke dynamics based user
authentication. In: 2016 IEEE 17th international symposium on computational intelligence and
informatics (CINTI), Budapest, Hungary, November 2016, pp 000105–000110. https://doi.org/
10.1109/CINTI.2016.7846387
9. Xie X, Xu Z (2020) Blind watermark detection based on K-S test in radio-frequency signals.
Electron Lett 56(1):30–32. https://doi.org/10.1049/el.2019.2478
10. Hu J, Yang W, Yang N, Zhou X, Cai Y (2016) On–off-based secure transmission design with
outdated channel state information. IEEE Trans Veh Technol 65(8):6075–6088. https://doi.org/
10.1109/TVT.2015.2477427
11. Hamamreh JM, Furqan HM, Ali Z, Sidhu GAS (2017) An efficient security method based on
exploiting channel state information (CSI). In: 2017 international conference on frontiers of
information technology (FIT), Islamabad, Pakistan, December 2017, pp 288–293. https://doi.
org/10.1109/FIT.2017.00058
12. Liau Q, Leow C, Ding Z (2016) Physical layer security using two-path successive relaying.
Sensors 16(6):846. https://doi.org/10.3390/s16060846
Security at the Application Layer and the Physical Layer in 5G 467
13. Efstathiou D (2019) A collaborative physical layer security scheme. Int J Electr Comput Eng
IJECE 9(3):1924. https://doi.org/10.11591/ijece.v9i3.pp1924-1934
14. Hojeij MR (2018) Resource allocation techniques for non-orthogonal multiple access systems,
p 190
Blockchain Based Security and Privacy
in VANETs
1 Introduction
VANET is the leading vehicular network used for communication because its assis-
tances in trouble-free driving as it can reduce traffic congestion and recurrence of
accidents. Traditionally, data regarding road traffic can only be collected through
cameras and surveys. On the contrary, VANET facilitates real-time data collection.
This data can be combined to get the real-time image of the road. Numerous applica-
tions have been developed through VANETs but they are lacking security and privacy
[1]. Therefore, there is a need to improve the privacy and security of VANETs.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 469
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_43
470 S. Iqbal et al.
Recently, blockchain has been adopted in various domains which can address
the basic security and privacy problems, especially for information reliability and
integrity as it is an immutable, transparent and irreversible database. It handles
program applications with the help of transaction trace reliably and conveniently.
Blockchain is a prominent and widely used application which is extensively
applicable in digital assets [2], distributed storage [3] and smart contracts [4].
Many surveys exist on blockchain [5, 6] that are addressing security and privacy
requirements in blockchain-based VANETs. These surveys emphasize on the appli-
cations and design. However, this research is based on the investigation of how
blockchain can be used to address various privacy and security requirements in
VANET.
Comparative study of various security and privacy requirements for VANET appli-
cations and blockchain based solutions are discussed. It emphasizes on constraints
and breaches of existing solutions and their impact on scalable scheme of blockchain
in VANETs.
This research work is arranged in 8 concise sections. Background of VANET and
applications are discussed in Sect. 2. Security and privacy requirements of VANET
are presented in Sect. 3. Section 4 discusses the constraints of existing VANET
solutions and their requirements. Section 5 explains the basic concepts of blockchain
and its usage. Section 6 discusses blockchain-based solutions addressing privacy and
security requirements of VANETs. Comparison of blockchain and non-blockchain
solutions is also presented in Sect. 6. Lastly, in Sect. 7 the paper is concluded.
2 Background
In this section emphases are made on the general concept of VANET and its appli-
cations while the next section discusses the fundamentals of blockchain and its
usage.
communication. V2I provide updates related to real time traffic and weather condi-
tion. Many applications are developed for VANETs which can be divided into below
mentioned three categories:
Vehicles in VANETs can create self-organizing systems without any potential infor-
mation of other vehicle. Consequently, the security and privacy requirements of
VANETs ought to be fulfilled before its deployment. Table 1 summarizes the VANET
applications with respect to these requirements that are discussed below:
Data Integrity. Keeping up the integrity of data implies that only authorized parties
can control the data legally. Attacks over the network can cause false modification
in information shared between nodes and it must be protected. This requirement will
also remain high in VANET applications.
Non-repudiation. It is the assurance of the data origin and integrity and does not
allow someone to deny from their act. Therefore, it is essential that information
transmission between nodes should be prevented from repudiation.
Data Verification. The data accumulated should be verified to guarantee that infor-
mation that is assembled is as precise as possible. Therefore, information must be
verified in VANET on regular basis to eliminate the fake data. Fake information
would mislead drivers that can lead to severe accidents or wastage of time and fuel
[12].
Mobility. Mobility is the capability to move freely. highly mobile nodes in VANET
entail less execution time of security protocols therefore learning based method
should be developed so that nodes can evolve to recognize each other’s behavior.
Mobility will be highly required in all applications so that vehicle can easily leave
or join the network for getting timely information [13].
Key Management. Key management is the set of operations that are fundamental
to create, maintain, secure, and control the usage of cryptographic keys. In VANETs,
key creation for allocation should be managed especially [13].
In the next Section various approaches are discussed that are dealing with the
above-mentioned requirements.
Different approaches for VANET security and privacy are presented below:
474 S. Iqbal et al.
Han et al. [20] proposed a protocol that handles efficiency and security problems.
It improves the efficiency of certification by using group based key broadcasting
technique without certification authority (CA) which an authorized vehicle from
being attacked using a deniable technique known as group key contract. However,
it is not a scalable scheme as it requires each vehicle to verify others throughout
the key contract resulting in a delay in verification. Conditional privacy scheme was
proposed by Xiaoyan Zhu [21] which is based on group signature. This scheme is
aimed at attaining conditional privacy through a hash chain technique which reduces
computational overhead. Group manager retains the key and every participant signs
the message on the behalf of the group. In this case, group signature verification is
more costly than traditional signature. Moreover, canceling compromised individuals
appropriately is still an issue. Kanchan and Chaudhari [22] presented a method based
on group signature combined with re- encryption method. In this method, a message is
broadcasted and is available to all the individuals of the group by using re-encryption
method. Nonetheless, privacy can be compromised in such method.
Blockchain Based Security and Privacy in VANETs 475
P. Asuquo et al. [23] discussed possible solutions for retaining location privacy based
on the use of cryptography for retaining the users. The approach is based on central-
ized method by which a central authority is accountable for forcing the secrecy
strategies. This authority can be the key dispersion framework inside a cryptographic
framework that is accountable for limiting the framework to the authentic clients.
Blockchain is a chain of records, called blocks, linked through cryptography. For each
block, a hash is generated with timestamp with transaction data of preceding block
[27]. It controls a dispersed record to permit transactions among nodes efficiently in
the network. It does not allow data modification and guarantees transaction integrity
[28]. It is a smart method of data transfer from A to B in completely safe and
mechanized way. Transaction process get started by one party that generates a block.
The block gets proved by millions of systems spread around the internet and is later
added to the chain that is kept across the network. This generates a record with unique
476 S. Iqbal et al.
history. Therefore, forging a record would mean that forging the whole chain, which
is impossible in such Scenario. The possible use cases of blockchain technology are
given below [29].
5.1 Currency
The first application that was introduced is known as Bitcoin. It is a digital currency
that allows safe transactions without central banking. Blockchain technology is
expected to use decentralization to advantage industries and developments where
trust level is required.
5.2 Finance
The utmost extensively known blockchain application lies in finance field, it guar-
antees much valued transparency among the transactional parties. Each transaction
in community, stocks, bonds or derivatives could be recorded in blocks and later be
confirmed by the authority for its legality. It’s simpler to identify money laundering
or extortion cases through stock trade moves.
Blockchain Based Security and Privacy in VANETs 477
5.3 Voting
5.5 Identity
Centralized schemes for VANETs cannot handle the increasing complexity of intel-
ligent transportation system (ITS) structures. The speedy evolution of Internet of
Vehicles (IoV) has provided large demanding situations for huge information storage,
478 S. Iqbal et al.
asset management, and information and transaction security. Recently, security and
privacy issues of VANETs are addressed through variety of blockchain mechanisms
within the literature few of them are discussed in this section. Z. Lu [30] proposed
a system based on public key structure to create a trust-based privacy preserving
model for VANETs. This study is focused on the methods based on privacy and
summarized three trust models to establish trust management in VANETs. N. Malik
[31] proposed a frame- work based on authentication using blockchain in VANETs
which covered only privacy concerns and ultimately failed to deal with security
of nodes communication. R. Shrestha et al. [32, 33] proposed a solution for nodes
communication security based on crypto trust point. However, they failed to address
privacy issues. A trust-based framework has been proposed by M. Singh [34] using
blockchain technology. It provides the trust environment based on proof of driving to
support information sharing and secure communication among vehicles. Lilli et al.
[35] provided an incentive announcement network based on blockchain technology.
It was proposed to guard vehicles’ privacy and encourage users to share traffic statis-
tics. However, location privacy is not considered in this model and the network
idleness to deliver data and to inform the RSI isn’t assessed. A blockchain-based
framework was proposed by A. Dorri [36] to report privacy and security of smart
vehicles. The research presents several cases related to remote program updates as
well as adaptable automotive insurance plans. They have neither given technical
details of their framework nor address the issues. X. Zhang [37] Emphasized the
extent of statistics generated via vehicular adhoc network and strain the significance
of Mobile Edge Computing (MEC). It was designed to compensate utilization of
resources in VANET based on blockchain. It helps in lowering the computational
overhead of blockchain. However, the MEC does no longer make it decentralized.
A secure system proposed by X. Zhang and X. Chen [38] for storage and data
sharing based on the blockchain consortium in VANETs. It can address the identity,
validity and message reliability issues but in result it generates additional overhead.
A secure key management framework based on blockchain has been proposed by A.
Lei [39] for the heterogeneous ITS systems. It gives good security in communica-
tion between vehicles and infrastructure. But it fails to conserve the vehicle’s privacy
which if exposed, can locate and put it at possible risk. A blockchain based decentral-
ized trust management system have been proposed by Z. Yang [40] in the vehicular
networks. In this, vehicle allot a positive vote to the neighbouring trusted vehicles
and upload the rating to the Roadside Units which are responsible for calculating
the trust score and set it in trusted blockchain. It has not provided vehicles’ vote
privacy mechanism and doesn’t consider the trustworthy weight of partaking vehi-
cles. A point-to-point trans- action model presented by J. Kang [41] is grounded on
the blockchain consortium that resolves the issue of scalability. It doesn’t completely
ensure the data exchange procedure. Trust point mechanism have been presented by
Singh and Kim [42] to secure the communication among vehicles. For data sharing
such methods couldn’t work in vehicle-based edge computing. It is because launching
a blockchain in resource limited vehicles is costly. Table 3 gives the summarize view
of the above discussed blockchain solutions with respect to VANET security and
privacy.
Blockchain Based Security and Privacy in VANETs 479
Table 3 Existing blockchain based solutions for security and privacy requirements
Year Paper reference Addressed security Addressed privacy requirements
requirements
2018 [30] No Anonymity
2017 [36] Data integrity Anonymity, Confidentiality
2018 [35] No Anonymity, Conditional privacy
2018 [31] Authentication No
2018 [32] Data verification, No
Authentication
2018 [33] Authentication, Data integrity No
2018 [34] Authenticity Anonymity
2018 [37] Data integrity No
2019 [38] Authentication, Data integrity, Anonymity
Availability, Data verification,
Non-repudiation
2017 [39] Key management No
2018 [40] Authentication, Data integrity No
2017 [41] Authentication, Data Conditional privacy
verification Data integrity, Key
management
2018 [42] Authentication, Availability No
7 Conclusion
This research has concluded that blockchain is one of the most innovative tech-
nologies that can address the various security and privacy requirements in VANET.
Initially, VANETs security and privacy requirements with respect to various applica-
tions are discussed. Secondly, existing security and privacy solutions for VANET are
presented. Further, this paper explained the integration and working of blockchain
with VANET and highlights how blockchain based solutions can address the secu-
rity and privacy requirements of VANET. Blockchain gives superior adaptability in
getting to the information and it gives better security and privacy in communication
between nodes. Lastly, this paper will be helpful in bringing up the studies related to
blockchain in VANET. This paper will provide fundamental thought to understand
the necessity of blockchain in VANET.
References
37. Zhang X, Li R, Cui B (2018) A security architecture of VANET based on blockchain and mobile
edge computing. In: 1st IEEE international conference on hot information-centric networking
(HotICN), Shenzhen, pp 258–259
38. Zhang X, Chen X (2019) Data security sharing and storage based on a consortium blockchain
in a vehicular adhoc network. IEEE Access 7:58241–58254
39. Lei A, Cruickshank H, Cao Y, Asuquo P, Ogah CPA, Sun Z (2017) Blockchainbased dynamic
key management for heterogeneous intelligent transportation systems. IEEE Internet Things J
4(6):1832–1843
40. Yang Z, Yang K, Lei L, Zheng K, Leung VCM (2018) Blockchain-based decentralized trust
management in vehicular networks. IEEE Internet Things J 6(2):1495–1505
41. Kang J, Yu R, Huang X, Maharjan S, Zhang Y, Hossain E (2017) Enabling localized peer-to-
peer electricity trading among plug-in hybrid electric vehicles using consortium blockchains.
IEEE Trans Industr Inf 13(6):3154–3164
42. Singh M, Kim S (2018) Trust Bit: reward-based intelligent vehicle commination using
blockchain paper. In: IEEE World Forum on Internet of Things (WF-IoT), Singapore, pp 62–67
Dynamic Data Protection for Mobile
Agents: XACML/ABAC Policies
Hassan Razouki
Abstract A mobile agent can move from one platform to another during runtime
to access data or resources. He moves with his own code and data. Agents take on
two roles in the runtime environment. First, agents are treated as subjects, because
agents access the different resources of a platform. Second, they are considered as
active resources in the execution environment. The protection of platform resources
against the attack of malicious mobile agents is already addressed in the literature.
However, the protection of code and agent data is an open and difficult issue due to
the fact that the execution environment has full control over the mobile agent.
A malicious platform may attempt to attack a mobile agent in order to analyze or
modify the data collected (Dynamic Data) by this agent while it is running on another
platform. To solve this problem, we have developed a security mechanism to manage
access to the agent’s dynamic data between the platforms visited by this agent. This
mechanism is implemented by the SAS (Security Adaptation System) which allows
defining the access control rules to the dynamic data of the mobile agent. The SAS
estimates the security level of each agent component (agent code in the form of
software components) and the degree of confidence of the platforms visited by this
agent. It then uses the multi-level model to define these rules. We adapt the ABAC
model and the XACML language to create and publish access control policies for
dynamic mobile agent data based on these rules.
1 Introduction
Most applications on the Internet require interaction between different entities across
the network to exchange data and distribute tasks. This increases the need for
autonomous entities capable of solving complex problems related to distributed
H. Razouki (B)
Faculty of Sciences and Technologies, University Sultan Moulay Slimane, Beni Mellal, Morocco
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 483
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_44
484 H. Razouki
systems. In this context, researchers are increasingly interested in the use of mobile
agent technology as it brings significant gains for several application areas, such
as Wireless Sensor Networks (WSN) [1], Cloud Computing [2], e-commerce [3],
information search [4], telecommunications [5], web services [6], transport system
optimization [7] and many other areas.
In particular, the safety issue is a barrier to the expansion of this technology [8,
9]. When an agent arrives on an execution platform, it exposes its code and data.
As a result, this platform will have the ability to modify them. It can modify the
agent’s data or the results obtained by the mobile agent within previous platforms
(dynamic data). The owner of an agent does not want an untrusted platform to be
able to modify the agent or the results he has obtained on other platforms [9]. In this
case, it is necessary to encrypt sensitive data in a way that only authorized platforms
can read or modify the data. The problem we encountered is how to secure access to
the agent’s encrypted dynamic data between the platforms visited by the agent.
To solve this problem, we have developed a mobile agent model based on a
software component [10]. Then, we associated with each agent a security policy
defined by the agent creator [11]. This policy contains a set of security requirements
that must be met by the visited platforms such as encryption and data signing. To
protect the sensitive dynamic data of the mobile agent, we thought to encrypt these
data by the visited platform (the encryption algorithm used is determined by the
agent’s security policy) [11], the encryption key is generated and stored by the KDC
server (Key Distribution Center).
To manage access to the decryption key between platforms, we have devel-
oped a Security Adaptation System (SAS) to create access control policies for the
agent’s encrypted dynamic data (it manages access to the decryption key). To create
these policies, the SAS estimates the security level of the mobile agent components
according to the security policy defined by the agent creator [11]. Then, it uses
access control constraints to create access control rules for agent data based on the
Bell and La Pudula multilevel model [12]. These rules are then transformed into
XACML/ABAC access control policies. These policies manage access to dynamic
data between the platforms visited by the agent.
The remainder of this document is organized as follows. In Sect. 2, we present the
architecture of our protection system. In Sect. 3, we present the estimated security
levels of the software components of a mobile agent and the confidence levels of
the platforms visited by the mobile agent to create the access control rules for the
dynamic data of this agent. Section 4, presents how to protect the agent’s dynamic
data during its execution. Section 5, presents the performance of this solution against
different types of attacks from a malicious platform. Finally, Sect. 6, concludes the
document and proposes directions for future work.
Dynamic Data Protection for Mobile Agents … 485
Our architecture is based on three phases: creation, publication, and use of access
control policies to protect mobile agent data. The interaction between the different
entities is illustrated in Fig. 1.
Our system consists of four components:
Server PAP (Policy Administration Point) allows us to create different XACML
policies for a mobile agent. The PAP stores the policies and makes them available
to the PDP. These policies represent the complete management policy that controls
the decisions made by the PDP.
PDP Server (Policy Decision Point) is the entity that manages authorization
requests from the KDC server, it is in charge of determining the rules and poli-
cies applicable to a given request. After evaluating the targets and conditions of
the ruleset, the PDP returns the response to the PDP. The PDP server relies on the
Certification Authority (CA) to verify the validity of the certificates of the different
entities.
Server LCA (Local Certificate Authority) is a component of PKI (Public Key
Infrastructure) that allows us to create and publish signed digital certificates
containing the public key and identity of each entity in our infrastructure. Digital
certificates are used to authenticate different entities in the system.
KDC (Key Distribution Center) is used by our system to protect agent data during
its execution. Therefore, only platforms authorized by the PDP server can access
encrypted mobile agent data. The KDC contains two components:
• Policy Enforcement Point (PEP) is a component that makes authorization requests
to the PDP, it enforces the authorization decision made by the PDP. It is the PEP
that technically performs the assignment or denial of symmetric keys to platforms
to encrypt its data or decrypt data from other platforms.
• Key Distribution Service (KDS) allows to intercept requests and verify the identity
of key requesters. The KDS uses a database to manage symmetric keys distributed
to platforms. The KDS component performs three functions: key creation, key
distribution, and key access control. It uses PEP to retrieve the authorization
decision from PDP.
In this section, we propose a new technique for creating access control rules for
mobile agent data access control using the Security Adaptation System (SAS). To
do so, we have estimated the security levels of the agent’s software components and
the confidence levels of the platforms visited by the agent.
We have adapted the multi-level model of Bell & La Padula (Bell and La Padula) to
create policies for controlling access to agent data. This security model is based on
the lattice concept that formalizes the security objectives and authorization scheme
of their policy, which can be used to demonstrate consistency.
Within the framework of a multi-level settlement, we have associated security
levels to the different components of the agent according to its security policy, and
degrees of confidence to the platforms running the agent. We defined the security
level of the components as follows:
Security Level of a Component (SLC) is the quadruplet (SR1, SR2, SR3, SR4)
of Security Requirements defined by the agent creator in the form of a security
policy. These requirements are classified according to four ordered security levels
(top secret, secret, confidential, non-confidential). The SLC reflects the sensitivity
of code and data produced by this component in a platform.
The principle of our approach is to assign a classification to each Security Require-
ment (SR). These requirements are derived from the agent’s security policy to protect
a particular component. There are four hierarchical security levels in order: top secret
(TS = 3), secret (S = 2), confidential (C = 1), and non-confidential (NC = 0).
The highest value of a requirement is considered the most secure. The four secu-
rity requirements are supported in the process of estimating the security level of a
Dynamic Data Protection for Mobile Agents … 487
functional component of the mobile agent. Table 1 below gives the classification of
security requirements for estimating the security level of a component.
Formally, we present the security level of a component as the set of security
requirements described in Table 1. These requirements are classified according to
the security policy defined by the agent creator. It is assumed that a mobile agent in
After the SAS has estimated the safety levels of the components, the next step is
to determine the degrees of confidence of the platforms visited by the agent.
Our mobile agent consists of a set of functional components and a security policy
to be followed during its execution. Each component runs on one or more platforms
depending on the agent itinerary. To manage access control to agent data between
platforms, the SAS defines a set of access control rules based on the security level
of the components and the degree of confidence of the visited platforms. We defined
the degree of confidence of a platform as follows:
Degree of confidence of a platform (DCP) is the quadruplet (DA1, DA2, DA3,
DA4) of Degrees of Authorization (DAi) estimated from one or more components
executed by that platform. Each Degree of Authorization (DAi) takes the maximum
value of the security requirement classification of the components executed by that
platform.
To maintain the confidentiality of data collected from execution platforms. An
agent contains n components (C) and is executed by m platforms (P). The SAS
analyses the agent’s security policy to estimate the security level of each component
(SLC). Then, it estimates the degree of confidence of each platform (DCP) from the
components executed by this platform. The DCP is expressed by Eq. 2:
⎧
⎪ ∀ j ∈ [1, m] ∃ A j ⊂ [1, n]/∀i ∈ A j the component Ci is run in P j
⎪
⎪
⎪
⎪ SLC i = (S R1i , S R2i , S R3i , S R4i )
⎪
⎨
DA1 j = max(SR1i ); DA2 j = max(SR2i )
i∈A j i∈A j (2)
⎪
⎪
⎪
⎪ DA3 = max (SR3 ); DA4 = max (SR4i )
⎪
⎪
j i j
⎩ i∈A j i∈A j
DCP j = DA1 j , DA2 j , DA3 j , DA4 j
At the end of this step, SAS estimates the degree of confidence of the different plat-
forms and then uses the multilevel model to create the rules for controlling access to
agent data. The security levels and degrees of confidence form a lattice that is partially
ordered by a dominance relationship denoted “≤”. The protection of data confiden-
tiality is expressed by the fact that a platform P has a DCP degree of confidence.
Dynamic Data Protection for Mobile Agents … 489
This platform can only access data collected by component C in another platform if
the degree of confidence of the platform (DCP) dominates the security level of this
component (SLC). There are three constraints for creating rules to control access to
its data:
Constraint 1: If a Pj platform is allowed to run a Ci component, then the Pj platform
has the right to add encrypted data to the agent after running this component.
Constraint 2: A Pj platform wants to read the data collected by a Ck component
only if the two following conditions are verified:
• The sum of all security requirements of component Ck must be less than or equal
to the sum of all platform authorization degrees Pj .
• The security requirement relates to the confidentiality of the data produced by the
component Ck must not be classified as “Top secret = 3”.
This constraint is expressed by Eq. 3:
⎧ 4
⎨
4
SRi (Ck) ≤ DAi (Pj)
(3)
⎩ i=1 i=1
SR4 (Ck) < 3
Finally, the SAS defines the rules for controlling access to agent data, based
on the degrees of confidence of the platforms and the security levels of the agent
components. Then, the SAS asks the PAP to create policies corresponding to these
rules. The PAP ensures that the access rules conform to XACML syntax and then
creates and stores the access control policies.
490 H. Razouki
Suppose that the first platform visited by the agent is PL-1, the agent route determines
the component that must be executed by this platform, as well as the security policy
used by the component [11]. After the agent authentication. Suppose the platform
wants to add data to the agent. In this case, the platform uses its Data Storage
Component (DSC) to add its data (XML format) to the running agent.
Assuming that the agent creator requires in its security policy the encryption of
the data produced by this component, the PL-1 platform uses the Access Control
Component (ACC) to send a request for a symmetric key to the KDS (The KDS is an
internal KDC server component). This request contains the identity of the platform,
the identity of the agent and the running component, the label of the data to be
encrypted, and the “Create” action. In our approach, we use three values for the
action attribute:
– Create: to request the creation of a new symmetric key to the KDS,
– Update and Read: are used to retrieve symmetric keys from KDS.
The platform’s request is signed by its private key. The KDS receives the request
and then authenticates the requestor of the key. Our solution is based on the access
control rules defined by the SAS at the time of agent adaptation. In this case, only
authorized platforms can add data to the agent. The KDS uses the PEP component to
send a SAML authorization request to the PDP, this request contains the necessary
attributes for decision making. The PDP determines the rules and policies applicable
to the PEP application. After evaluating the targets and conditions of the set of rules,
the PDP returns the authorization decision to the PDP. If the response is favorable
“Permit”, the PDP gives authorization to the KDS to generate a new symmetric key
specific to the PL-1 platform. The KDS stores the key in an array indexed by platform
identity. After registering the key, the KDS sends the generated key to the ACC, the
Data Storage Component (DSC) uses the key to encrypt this data. The used key is
then destroyed before the agent migrates to a new platform. Note that the key is later
used by the KDS to provide keys to authorized platforms to decrypt this data.
When the agent migrates from one platform to another, it moves with it the data of
its execution in the previous platforms, we take the example illustrated in Fig. 2.
The platform (PL-2) wishes to read or update the data created and encrypted by the
platform (PL-1).
After the migration of the agent from platform PL-1 to platform PL-2 (step 1).
The PL-2 platform authenticates the agent via AS and LCA servers (step 2). Then
Dynamic Data Protection for Mobile Agents … 491
it requests the symmetric key from KDS to decrypt the data encrypted by PL-1
(steps 3 and 4). This request contains the identity of the platform, the agent and
the component that produces the data in the PL-1 platform, the data label and the
action to be performed “Update” or “Read”. The KDS verifies the authenticity and
integrity of the request (step 5) and then redirects the request to the PDP. The PDP
then sends a request for SAML authorization to the PDP with the necessary attributes
for decision making (step 6.1). The PDP evaluates the rules and policies applicable
to the application (Step 6.2) and sends its decision to the PDP (Step 6.3). If the
decision is favorable, the PDP asks the KDS to search and send the symmetric key
to the platform (steps 7.1 and 7.2). The CCA retrieves the symmetric key and then
decrypts the data encrypted by the PL-1 platform (step 8). This illustration shows how
one platform decrypts data encrypted by other platforms, as well as the interaction
between the different entities to recover the symmetric key.
In this section, we will discuss the security mechanisms used against different types
of attacks on dynamic mobile agent data.
• Dynamic data spying: Agent security policy requires encryption of data collected
by a component during its execution. In this case, the execution platform sends an
asymmetric key request to the KDC. If the platform is allowed to add the encrypted
data to the agent, the KDC generates a new key specific to that platform. The
platform uses this key to encrypt its data with the XML-Encryption specification.
Note that we used XML to store this data in the mobile agent.
492 H. Razouki
• Alteration of dynamic data: The mobile agent requires in its security policy the
signature of the data produced by certain components. In this case, platforms use
their public key certificates to sign this data. This signature is done through the
XML-Signature specification.
• Dynamic data access control of the mobile agent: the mobile agent migrates
from one platform to another with encrypted data. To manage access control
to this data between platforms, we have created XACML/ABAC access control
policies. These policies are based on the rules defined by SAS at the time of agent
adaptation. With this security mechanism, only platforms authorized by the PDP
can retrieve the symmetric key from KDC to access this encrypted data.
6 Conclusion
References
1. Mehak, Khandnor P (2019) Energy efficient data aggregation using multiple mobile agents
in wireless sensor network. In: Advances in intelligent systems and computing, vol 669, pp
279–287
2. Aditya Sai Srinivas T, Ramasubbareddy S, Govinda K (2019) Discovery of web services using
mobile agents in cloud environment. In: Lecture notes in networks and systems, vol 74, pp
465–471
3. Roopa GM, Nirmala CR (2019) Intelligent mobile agent framework for searching popular
e-advertisements. In: Advances in intelligent systems and computing, vol 714, pp 177–189
4. Mwiti JI, Kahonge M, Abande E (2019) Mobile agents in distributed environment for
information retrieval. Int J Appl Inf Syst 12(20):8–10
5. Kang H, Kim H, Kwon YM (2019) RECEN: resilient MANET based centralized multi
robot system using mobile agent system. In: 2019 IEEE symposium series on computational
intelligence (SSCI), pp 1952–1958
6. Seghier NB, Kazar O (2020) Mobile agent and ontology approach for web service discovery
using QoS. Int J Reason-Based Intell Syst 12(1):17–33
7. Shehada D, Yeun CY, Jamal Zemerly M, Al-Qutayri M, Hammadi YAl (2017) Secure mobile
agent protocol for vehicular communication systems in smart cities. In: Information innovation
technology in smart cities, pp 251–271
8. Razouki H, Hair A (2014) Self-adaptive security for mobiles agents. Int J Comput Appl 94(13)
9. Razouki H, Hair A (2015) Security for mobile agents: trust estimate for platforms. TELKOM-
NIKA Indonesian J Electr Eng 15(2):381–389
Dynamic Data Protection for Mobile Agents … 493
10. Razouki H, Hair A (2013) Components-based software architecture for secure mobile agents
via two strategies of adaptation. In: 2013 8th international conference on intelligent systems:
theories and applications, SITA 2013
11. Razouki H (2019) Security policy modelling in the mobile agent system. Int J Comput Netw
Inf Secur (IJCNIS) 11(10):26–36
12. Bell DE, La Padula LJ (1976) Secure computer system: unified exposition and multics
interpretation
Securing the Storage of Passwords Based
on the MD5 HASH Transformation
Abstract The hash function is designed to fold an input array of any size into a bit
string. For the MD5, this is a very convenient algorithm for comparison and storage.
For example, Unix systems use it to store user passwords in a 128-bit encrypted
format. It is widely used to check the integrity of files. The length of the output string
is 128 bits. When it first appeared, it was impossible to break it, but today this is
no longer the case. Many programs select the source word based on the hash. The
vast majority of them make an enumeration in the dictionary. However, Rainbow
Crack methods are based on generating a set of hashes from a set of characters to
search for a hash using the resulting base. Besides, the MD5, like any hash function,
has a concept such as collisions - it is about getting the same hashes for different
characters’ source strings. This paper will propose a method capable of increasing
password storage security based on the random generation of an MD5 HASH and
then use it to transform the original HASH according to pre-defined rules before
storing it in the database.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 495
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_45
496 H. Touil et al.
complexity, bit depth, and other properties. Today, hash has acquired many new
properties and is used in various fields of information technology. The main require-
ment for hash functions is the uniform distribution of their values for a random
selection of argument values. For cryptographic hash functions, it is also crucial
that the slightest change in the argument changes the function a lot. It is called the
avalanche effect [1–4].
The impossibility of fabrication requires that it is complicated to find a message
with the correct bending value. The second is the incredible complexity of matching
a given message with a known bend value with another message with the right bend
value [5–7].
Unidirectionality is understood as the incredible complexity of finding a message
by a given convolution value. It should be noted that there are currently no hash
functions used with proven unidirectionality.
Collision resistance refers to the difficulty of finding a pair of messages with
the same convolution values. Usually, the discovery by cryptanalysts of a collision
construction method serves as a first signal that the algorithm is obsolete and the
need to replace it soon [8–14].
The resistance to finding the second pre-image is understood as the complexity
of finding a second message with the same convolution value for a given message
with a known convolution value. There are two critical types of cryptographic hash
functions: key and keyless. Key hash functions are called message authentication
codes. They can be used, without additional means, to ensure both the data source’s
accuracy and data integrity in systems with trusted users. Keyless hash functions are
called error detection codes. They enable additional means (such as encryption) to
ensure data integrity. These hash functions can be used in systems with trusted and
untrusted users [15–21]. In the rest of this paper, we will see the works in relation.
Then we will dissect our method ending with the experiments.
2 Related Works
A set of studies has been proposed to ensure password storage security: the technique
[22] capable of hiding user data and making them abstract and can also support all
forgery attacks, using a secure remote authentication scheme based on smart cards.
Kumar [23] et al. Presents a new advanced security architecture for user identifica-
tion, which includes two authentication factors, encryption, and decryption of data
uploaded to the cloud using an AES system, verification and locking of users by
the administrator, recovery of user IP data, distributed database storage, i.e., data is
stored in tiers, which means that user login data is stored in one database and encryp-
tion/decryption data (downloaded files, key) is stored in different databases. Using
the hash technique, Matsunaga [24] et al. uses a message authentication protocol
using a homo-morphic hash function without full key security in a WANET envi-
ronment with limited resources for each node. Bebe [25] et al. invent a mechanism
under the name "OSA-SHSDS" to store user data in the cloud with a higher privacy
Securing the Storage of Passwords Based on the MD5 HASH Transformation 497
rate and less complicated space. In order to store the data in the server, the cloud
user’s data must be recorded. After registration, the cloud server generates the ID
and password for each registered cloud user. Each time the user wants to store the
data, he has to log in to the cloud server with his ID and password to authenticate.
3 Proposed Method
Our method consists of two different phases. In the first phase, the “subscription
phase,” where the registration page is invoked, the user fills in and sends the form.
Then the server extracts the submitted information, username, password, and other
additional information. The most interesting for us is the password that will be
hashed by the MD5 hash algorithm. The server will then generate a second hash
(MD5) randomly, as explained (Fig. 1).
The next step is to build the new hash according to the algorithm (Fig. 3). The
new HASH will be stored in the database, and then the end-user will receive the two
OUTPUTs first and then size so that they can use it during the next authentication
“authentication phase” (Fig. 2).
Database
4
Step 5: verification phase, ap-
ply the reverse process and
make a comparison between the
Step 1: Authentication
generated HASH and the initial
process
HASH.
Database
A. Pre-image
A pre-image attack on cryptographic hash functions attempts to find a message that
has a specific hash value. A cryptographic hash function must resist attacks on its
pre-image (set of possible entries).
With an ordinary MD5 hash, one can identify:
L: hash size estimated in 128bits.
N: Number of possible hash such that N = 2L.
P: the probability of finding the right HASH such that: P = 1/N.
Then the number of trials will be identified as follows:
∞
E(N) = i*(1 − p)(i−1) * P
i= 0
With our method, the complexity of falling on the HASH corresponds to the
password is almost impossible, given the significant number of tests the hacker must
perform.
B. Collision attack √
If a hash function produces N output bits, an attacker that only computes 2N hash
operations on a random input is likely to find two matching outputs. If there is a more
straightforward method than this brute force attack, it is generally considered a flaw
in the MD5 hash function.
Using our method, this hypothesis will not be beneficial to detect the password
given the change in hash size plus the reasonably large number of iterations.
Securing the Storage of Passwords Based on the MD5 HASH Transformation 499
4 Experimentation
2) Generate a 128 bits MD5 HASH randomly, based on a random integer, for
example: “1425”.
3) As described in the algorithm (Fig. 3), the two parameters first and size must
be generated as follows:
In this case, we will work directly on the hash in hexadecimal and will not need
to convert it to binary.
The goal is to build a new 256-bit HASH from the above two HASHs by alternating
construction starting with the Original HASH.
The goal is to build a new 256-bit HASH from the two HASHes above by
alternating construction starting with the random HASH.
Later this hash will be stored in the database with a size of 256 bits.
We can notice through these two examples that the generated HASHs are different.
Securing the Storage of Passwords Based on the MD5 HASH Transformation 501
The reverse operation capable of distinguishing the original HASH from the random
one, taking the HASH generated on the second example.
5 Conclusion
We have demonstrated a new authentication security specialist method that will make
pre-image and collision attacks almost impossible, given the new HASH size. Our
method offers a reasonably high level of security. Even if a hacker manages to access
our database and decrypt the HASH, he will inevitably come across a value that
has nothing to do with the password stored with a good user experience and very
low delays. The security of passwords has always been an essential concern in the
deployment of platforms, applications. However, it is almost impossible for a hacker
to find the secret being used with our approach, given the number of possibilities we
have and the new size of the hash. Moreover, even if we reuse the same password in
multiple websites, this will pose any security problems since each apparent hash is
unique.
502 H. Touil et al.
References
1. Touil H, El Akkad N, Satori K (2020) Text Encryption: Hybrid cryptographic method using
Vigenere and Hill Ciphers. In: 2020 International Conference on Intelligent Systems and
Computer Vision (ISCV), Fez, Morocco, pp. 1–6
2. Touil H, El Akkad N, Satori K (2020) H-Rotation: Secure storage and retrieval of passphrases
on the authentication process. Int J Safety Security Eng 10(6):785–796
3. El Akkad N, El Hazzat S, Saaidi A, Satori K (2016) Reconstruction of 3D scenes by camera
self-calibration and using genetic algorithms. 3D Research, 6(7): 1–17
4. Elazzaby F, El Akkad N, Kabbaj S (2019) A new encryption approach based on four squares
and Zigzag. The 1st International Conference on Embedded Systems and Artificial Intelligence,
ESAI
5. Es-sabry M, El akkad N, Merras M, Saaidi A, Satori K (2019) A new color image encryption
using random numbers generation and linear functions. In: The 1st International Conference
on Embedded Systems and Artificial Intelligence, ESAI
6. El hazzat S, Merras M, El akkad N, Saaidi A, Satori K (2019) Enhancement of sparse 3D
reconstruction using a modified match propagation based on particle swarm optimization.
Multimedia Tools and Applications (Springer), 78, pp. 14251–14276
7. Cheng R, Jin Y (2015) A competitive swarm optimizer for large scale optimization. IEEE Trans
Cybern 45(2), art. no. 6819057, pp. 191–204
8. Zhang W-X, Chen W-N, Zhang JA (2016) Dynamic competitive swarm optimizer based-on
entropy for large scale optimization. In: Proceedings of the 8th International Conference on
Advanced Computational Intelligence, ICACI 2016, art. no. 7449853, pp 365–371
9. Al-Tashi Q, Abdul Kadir SJ, Rais HM, Mirjalili S, Alhussian H (2019) Binary Optimization
Using Hybrid Grey Wolf Optimization for Feature Selection (Open Access) (2019) IEEE
Access, 7, art. no. 8672550, pp. 39496–39508
10. Singh H, Sharma M, Sharma S, Anand P (2019) A New Hybrid Algorithm Based on Grey
Wolf Optimization and Crow Search Algorithm for Unconstrained Function Optimization and
Feature Selection (Open Access) IEEE Access, 7, art. no. 8643348, pp 26343–26361
11. Mohapatra P, Nath Das K, Roy S (2017) A modified competitive swarm optimizer for large
scale optimization problems. Appl Soft Comput J 59:340–362
12. Das S, De S, Kumar R (2016) Implementation and comparative analysis of RSA and MD5
algorithm. Int J Comput Appl (0975–8887), 141(9)
13. Hossain A, Islam’S K, Kumar S, Nashiry M (2012) Cryptanalyzing of message digest
algorithms MD4 and MD5. Int J Cryptography and Inf Security, 2(1)
14. Niharika M, Ethiraj R (2015) Design and implementation of hybrid RC4 and Sha-2 algorithms
for WIFI applications. Int J Emerg Eng Res Technol
15. El Akkad N, Merras M, Baataoui A, Saaidi A, Satori K (2018) Camera self-calibration having
the varying parameters and based on homography of the plane at infinity. Multimedia Tools
Appl 77(11):14055–14075
16. El Akkad NE, Merras M, Saaidi A, Satori K (2013) Robust method for self-calibration of
cameras having the varying intrinsic parameters. J Theor Appl Inf Technol 50(1):57–67
17. El Akkad NE, Merras M, Saaidi A, Satori K (2013) Camera self-calibration with varying
parameters from two views. WSEAS Trans Inf Sci Appl 10(11):356–367
18. El Akkad N, Saaidi A, Satori K (2012) Self-calibration based on a circle of the cameras
having the varying intrinsic parameters. In: Proceedings of 2012 International Conference on
Multimedia Computing and Systems, ICMCS, pp 161–166
19. Merras M, El Akkad N, Saaidi A, Nazih AG, Satori K (2014) Camera calibration with varying
parameters based on improved genetic algorithm. WSEAS Trans Comput 13:129–137
20. El Hazzat S, Merras M, El Akkad N, Saaidi A, Satori K (2018) 3D reconstruction system based
on incremental structure from motion using a camera with varying parameters. Visual Comput
34(10):1443–1460
21. Merras M, Saaidi A, El Akkad N, Satori K (2018) Multi-view 3D reconstruction and modeling
of the unknown 3D scenes using genetic algorithms. Soft Comput 22(19):6271–6289
Securing the Storage of Passwords Based on the MD5 HASH Transformation 503
22. Mukesh S, Tejas P, Anuj J (2020) Security analysis on remote user authentication methods:
Lecture Notes on Data Engineering and Communications Technologies, vol 31, pp 506–513
23. Kumar S, Jafri SAA, Nigam N, Gupta N, Gupta G, Singh SK (2019) A new user identity based
authentication, using security and distributed for cloud computing: IOP Conference Series:
Materials Science and Engineering, vol 748, Issue 1, Article number 012026. In: International
Conference on Mechanical and Energy Technologies 2019, ICMET 2019; Galgotias College
of Engineering and Technology, Greater Noida, Uttar Pradesh, India
24. Matsunaga A (2019) Message authentication scheme for ad hoc Networks with homomorphic
hash function. In: Proceedings of 2019 2nd World Symposium on Communication Engineering,
WSCE 2019 December 2019, Article number 9040951, pp 138–141
25. Bebe PC, Akila D (2019) Orchini similarity user authentication based streebog hash function
for secured data storage in cloud. In: Proceedings of 2019 International Conference on Compu-
tational Intelligence and Knowledge Economy, ICCIKE 2019, December 2019, Article number
9004393, pp 461–466
Is Blockchain Technology Applicable
in Small and Medium-Sized Enterprises?
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 505
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_46
506 Z. Chabani et al.
1 Introduction
Managing human resources in SMEs is challenging, considering their size and finan-
cial capacity. Compared to large-sized companies, SMEs cannot afford to have a
large team of human resource specialists. Generally, small firms tend to rely on
informal human resource management processes [6]. They do not have formal Human
Resource Management structures and policies. In addition, they often do not have
formal training for managers in the implementation of HRM practices.
According to some authors, blockchain technology helps companies become
stronger competitors when applied to different organizational functions. HRM is one
of these organizational functions that can be revolutionized and optimized, thanks
Is Blockchain Technology Applicable … 507
to blockchain technology. In this context, it may be relevant for SMEs to use this
technology in order to optimize the efficiency of their HRM.
The decentralized nature of blockchain should help automate information veri-
fication, enhancing the accuracy of staffing, managing, and reducing costs. It helps
organizations optimize the human resource department’s administrative task, record,
search, and validate employees’ data without the intervention of a mediator, upon
the condition of having the employee’s agreement [18].
In addition to eliminating the involvement of a third party, the blockchain
technology’s main contributions are privacy, security and transparency. The smart
contract built-in blockchain technology helps organizations store encrypted data in
a decentralized location, increasing the level of privacy and security. Furthermore,
blockchain technology does not limit the access of information to managers; it is
accessible to the company’s employees [18]. In this context, they can access the
information related to HRM functions within the organization, like training, perfor-
mance appraisal, and recruitment. This open access to employees fosters transparency
in the organization, increases the perception of fairness, improves the company’s
credibility, and develops a healthy and positive atmosphere in the organization [18].
Blockchain technology may result in more accurate methods of hiring employees.
It can be a source of significant potential solutions to eliminate recruitment issues by
supporting decision-making related to recruitment. Blockchain technology can help
organizations better recruit specialized talents by obtaining better information about
applicants, which cannot get with the traditional recruitment methods. Organizations
can obtain trustworthy information no matter the country or region the applicant
has obtained his degree or employment experience. With the employee’s consent,
employers can have access and can check the veracity of the provided information.
Besides staffing, blockchain technology should have a positive impact on compen-
sation, mainly employees’ payroll system. It makes it easier, shorter and faster to
process without the mediating role of banks. Thanks to smart contracts, automatic
payment is possible, based on a specific code and a specific persistent condition
related to the payroll system, which reduces significatively the risk of fraud.
Blockchain technology can also have a positive impact on employees’ training and
development as well as performance management; it helps to store information about
training and employees’ movements and promotions within the organization, with
access that is not limited to managers, which will create transparency considering
that the information is also available to employees [18].
Therefore, blockchain technology may optimize the HRM functions in SMEs.
However, some authors argue that the implementation of blockchain technology
should be adapted to SMEs’ business needs and should follow a tailored approach
when it is adopted due to some disadvantages. For instance; The DLT can store
tamper-proof information and other details about an employee, which is otherwise not
needed in the company, and exposes the information to the peers. Using blockchain
presents privacy and confidentiality challenges on the personal information of specific
employees [13]. In some cases, personal information is required for privacy protec-
tion and then hacked, thus, exposing the employees or senior managers to chal-
lenging privacy issues and potential fraud. The other challenge concerns an increase
508 Z. Chabani et al.
at decisions on managerial changes for the DLT structure is often difficult. Tradi-
tionally, the regulators of the financial sectors have relied on effective centralized
management for their decision. As indicated by Gatteschi et al. [5], the administra-
tors may be subject to specific DLT arrangements that may be difficult to enforce.
Thus, SMEs need to be aware of this potential difficulty.
Transition to using blockchain also brings about legal and regulatory challenges.
The developments of standards at the industrial level are necessary, yet they are still
at the incubation stage [10]. This implies that for a single SME to adopt blockchain
when industrial standards are not set will bring complexity in business operations
[3]. In addition, legal clarity regarding the jurisdiction and ownership of the DLT
system may be challenging. The primary reason is that there is no legal entity that
controls distributed ledger in the blockchain. The fact that the distributed ledgers in
blockchain are immutable creates a gap regarding how transaction disputes can be
resolved as they arise [15].
Financial resources represent another challenge for SMEs, often struggling with how
to optimize their financial resources, capacity, and management.
Based on blockchain technology, the DLT is known for having the capacity to
enable direct transmission of token or digital values between two parties and a
decentralized record-keeping. Resultantly, there is no need for a central authority
that controls the ledger. The results are; better scalability; they reach the market
faster and at a reduced cost. The other advantage is that DLT enhances transparency
and makes auditing easier because all the network members retain a full copy of
the ledger (Encrypted so as not to make changes). Resultantly, given that all ledgers
are encrypted, and the absence of central authority makes it hard to commit fraud.
At the same time, it eliminates reconciliation costs. The DLT transaction approach
features transactional exchanges on a peer to peer basis with conditions (such as the
date of release and rules for the failure of loan repayment) that are embedded and
predetermined [12]. Consequently, there is no need to set up additional business rules
for collaterals between transacting parties.
Besides, the blockchain can offer an audit trail that is immutable and verifiable
for any form of digital and physical assets. The DLT also removes friction in trans-
actions, thus enhancing efficiency and speed. Besides, DLT ensures that friction in
the clearing of settlements and the transactions is removed, increasing efficiency in
the accounting sector. The process is fully automated, and the intermediaries are
eliminated to ensure that the settlement process is cleared within the shortest time
[12]. The other advantage is that when well secured, the DLT offers more resilience
against cyber-attacks than the traditional system. The reason is that with DLT, a single
510 Z. Chabani et al.
point of security is eliminated and replaced with multiple security points; hence it
will be challenging for attackers to eliminate all points.
One of the disadvantages pertains to blockchain’s scalability, both with regard to
the speed of verification and the volume of transactions [5]. The transaction speed of
blockchains that do not require permissions is limited. For example, in Bitcoin, the
DLT has been designed to only allow 4–7 transactions per second due to the block
size limit to a single megabyte. This disadvantage can be resolved by increasing the
transaction throughput, as is the case with Ethereum [9]. Several factors must be
considered when appropriating the blockchains to SMEs, given that one model may
be appropriate for a particular business and, at the same time, disadvantageous to the
other.
The integration of DLT in financial infrastructure can be very complicated. Specif-
ically, the integration requires coordination from the entire industry. Coordination
and collaboration at an industry level may require additional expenses, especially
at the initial implementation stages. The other disadvantage is that when there is a
single weakness in the blockchain, the attackers can capitalize on committing fraud.
The success of DLT thus relies on the nature of blockchain and the measures taken
to ensure that the algorithms determining the core values of the DLT system are
not compromised. However, it is possible to maintain integrity and complexities
that make it difficult for attackers to find any fault line. For instance, the Bitcoin
blockchain has never been attacked since it adopted the DLT system in 2009.
This literature review has shown that blockchain provides SMEs an opportunity to
eliminate the intermediaries within the company. Specifically, blockchain thinking
gives a business a chance to experience accountability, the Internet of Things (IoT),
decentralized decision-making, and sustainability within an industry. All the advan-
tages and disadvantages are summarized in Table 1 to prepare different managerial
functions, including strategic management, HRM, and finance, to understand the
positive and negative effects of adopting DLT.
Scalability of the DLT System. It is essential for the DLT to have an appropriate
scale that will ensure it is not vulnerable to cyber-attacks. The DLT should have
flexible governance and regulations so that it is easily upgradeable. This will also
ensure that a hard fork’s chances are significantly lowered because the regulations
are easy to follow. The blockchain should also be designed to ensure interoperability
in applications of the IoT and smart grid. The blockchain market should not just be
efficient but also transparent to build trust for all stakeholders in the chain. The DLT
Is Blockchain Technology Applicable … 511
Table 1 The advantages and disadvantages of adopting the blockchain technology in the SMEs.
The perspective Advantages Disadvantages
Human resource management To promote transparency in the Privacy and confidentiality
organization problems
To increase the perception of
fairness within the
organization
To improve the company’s Blockchain adoption may
credibility cause a reshuffling of job
To develop a healthy and positions
positive atmosphere in the
organization
To improve compensation
management, mainly
employees’ payroll system
To optimize the management
of training, employees’
movements and promotions
Strategic management To facilitate the market Applying blockchain may
analysis for better necessitate changing the whole
decision-making structure of the organization to
Proper valuation of the fit the new technology
company’s assets
The activities run faster than in Standards and regulations at
the traditional forms of the industrial level are not clear
organizations
Planning goals and achieving
them is more comfortable than
in the traditional form of
organizations
Finance To Enhance transparency and The initial stage of adopting
making auditing easier blockchain requires a
considerable amount of capital
To eliminate reconciliation Cyber-attacks, if successful,
costs can lead to substantial financial
losses
is fast evolving to meet these objectives and ensure that ledgers and data are securely
recorded across multiple sources.
Interoperability of DLT. It is essential that the DLT design that has been adopted
by an SME can be migrated to a different blockchain whenever the need arises.
For instance, the SME may purchase a new subsidiary and transfer the current DLT
to the new premise without experiencing complications of any kind. Besides, in
consideration of the DLT interoperability, it is crucial to ensure that the migration
of the DLT is completed at a minimal cost. There is a shift in focus from a single
512 Z. Chabani et al.
blockchain, such as Bitcoin, to multiple chains that interoperate together and offer
niche services to all the users.
Privacy by Design. The issue of privacy was predominant in most literature included
in this study. It emerged that blockchain systems that have not been designed to
protect users’ data are more susceptible to unauthorized access and cyber-attacks.
DLT designers should ensure that measures are taken to limit the collection of
personal information. In cases where it is necessary to collect confidential infor-
mation, it is necessary to protect it. Findings from a study done by Islam and Hasan
[8], revealed that the DLTs in energy training is fond of using pseudonyms to protect
personal data. However, the pseudonyms can be misleading since it can only be valid
as long as the pseudonym has not been linked to a person. However, the moment a
connection is linked between a person and the pseudonym reveals all the confiden-
tial information. The other shortcoming of the use of pseudonyms is that it prevents
enterprises from sharing propriety information in a secure manner [19]. SMEs can
adopt privacy-enhancing information at the conception stage of the DLT system.
6 Conclusion
References
19. Wu H, Li Z, King B, Ben Miled Z, Wassick J, Tazelaar J (2017) A distributed ledger for
supply chain physical distribution visibility. Information 8(4):137. https://doi.org/10.3390/inf
o8040137
20. Yaga D, Mell P, Roby N, Scarfone K (2018) Blockchain technology overview. National Institute
of Standards and Technology, Gaithersburg, p 6. https://doi.org/10.6028/NIST.IR.8202
Anomaly Detection from Log Files Using
Multidimensional Analysis Model
1 Introduction
The global volume of digital data production is growing at a fast pace [1, 2], and
IT security has become increasingly important for businesses and organizations that
constantly store and transfer a huge pile of critical data across their local or wide
area networks [3, 4].
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 515
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_47
516 Y. Azizi et al.
Servers, firewalls, and other IT equipment keep updated log files where are record
all events and transactions. These files retrace the history of events occurring on a
server or application. The exploitation of logs represents a relevant operation to
improve security measures [5–7].
In this paper, we present our decision-making approach for log file exploitation
based on data warehousing techniques. We collect data from log files, which represent
a mine of information to exploit, we save them in a data warehouse to facilitate navi-
gation and analysis to produce a useful knowledge. This methodology can provide
a reliable view of the security level of our information system. In addition, it aims
to detect potential computer threats and enables the most appropriate correction or
reinforcement measures to be adopted accordingly.
This paper is organized as follows: Section 2 presents main related works that
have exploited log files for extracting useful information. Then, Section 3 illustrates
our methodology that we use to process log files to extract data on attack attempts,
intrusions and performance indicators. Before concluding, in the last section, we
show a case study on web server logs.
2 Related Work
In order to guarantee the security of the systems, it has become vital to develop new
monitoring and anomaly detection solutions, in this context log files have become
a very useful source of data. Anomaly detection through logs is a problem that has
already been covered in the literature [8–10], using a wide variety of analysis and
exploitation models based on different approaches. In our previous work [11, 12],
we have provided a methodology of security analysis that aims to apply Big Data
techniques, such as MapReduce, over several system log files in order to locate and
extract data probably related to attacks made by malicious users whose intends to
compromise a system.
In [13], Authors present the methodology used in pre-processing of high volume
web log files, studying the statics of website and learning the user behavior using
Apache Flume and Pig.
In [14], the researchers propose an automated tool to support developer investi-
gation of performance issues in distributed systems. They aim to leverage the vast
log data available from large scale systems, while reducing the level of knowledge
required for a developer to use their tool uses machine learning techniques. In [15],
authors exploit data mining techniques to identify students’ online SRL. They exam-
ined both self-reported self-regulation surveys and log files to predict online students’
achievements.
Anomaly Detection from Log Files … 517
Business Intelligence (BI) encompasses all the means, methods, tools for collecting,
consolidating, modeling, and restoring data from internal systems and external
sources, in order to offer decision support to professionals and decisions-makers.
In the literature, several recent works are based on the use of BI approaches and
data warehousing techniques, in [16], authors propose a data warehouse model for
climate big data enables the identification of weather patterns that would be useful
for agriculture fields, climatic change studies and contingency plans over weather
extreme conditions, in [17], they used data warehouse model for the analysis of field
and laboratory data from multiple heterogeneous photovoltaic test sites.
The objective of Business intelligence is to improve the timeliness and quality of
information, and enable managers to be able to better understand the actual state to
deliver valuable information for making a decision. BI is an area of Decision Support
System (DSS), the purpose of which is to gather and analyze past, present and future
data in order to facilitate monitoring and detailed understanding of the environment,
results, internal and external actors, to optimize strategic decisions and the conduct
of operations.
The architecture of BI can be containing three essential phases:
• Data Capture/Acquisition: operating data must be collected and recovered from
the various available sources, filtered and adapted in view of a decision-making
use,
• Data Storage: the data is structured, centralized and made available in the data
warehouse
• Data Access & Analysis: the end user must be able to analyze the information
and draw conclusions
4 Proposed Methodology
We have been witnessing for a few years now a quantitative explosion of data that
has illustrated three major characteristics of data: volume, variety and velocity. This
has exponentially increased the complexity of analysis and processing.
In this sense, we propose a methodology based on the decision-making infor-
mation system using data warehouse techniques for integrating, coordinating and
analysis this immense diversity of log files (Fig. 1).
Our proposed methodology includes the following layers:
• Collection layer: The collection is done from log files (error logs, access logs …).
These files can be in different formats and represent different systems. These files
are therefore generally heterogeneous, which is why we will have to go through
a so-called integration phase to be able to manipulate them before storing them
in our decision support system.
518 Y. Azizi et al.
• ETL layer: It is at this level that the first software layer of the decision-making
environment, namely the ETL (Extract, transform, loading). This layer offers
functions for extracting all events from different log file sources, transforming
these data (homogenization, filtering, calculation) and loading them into the DW
(data warehouse). It guarantees the relocation of the computing load and a better
availability of sources.
• Data Warehouse Layer: allows to store the data in a Data warehouse, the main
features of this data warehouse are subject-oriented, integrate, time-variant and
non-volatile.
• Data mart layer: a Datamart is a sub-element of the data warehouse, collects a set
of organized, targeted, aggregated and aggregated data to meet specific needs.
• OLAP cube Layer: multidimensional analysis is the ability to analyze data that
has been aggregated into several dimensions. So we wanted to access data already
aggregated according to the needs of the user, quickly and efficiently. For this we
use OLAP hypercube, the data are represented in n-dimensional hypercubes, and
structured along several analysis axes (dimensions).
• Reporting/analysis Layer: The last phase concerns the restitution of the results
The main objective of our proposed methodology is the exploitation and control of
log files to extract knowledge related to system security and also to determine and
control key performance indicators to make better decisions.
In this case study, we exploited log files of web servers as a source of data, these
files can contain verbose details about the inner workings of our servers.
There are two types of log: access log and error log [18]:
• Access Log records all requests that were made of this server including the client
IP address, URL, response code, response size etc.
• Error Log records all the details such as Timestamp, Severity, Application name,
Error message ID, Error message details (Fig. 2).
Anomaly Detection from Log Files … 519
The data warehouse modeling can follow one of the multidimensional data building
models, namely the star model, the snowflake model or the constellation model.
These models involve a table of facts as well as tables of dimensions.
We opted for the star model, which is the classic model. It is considered to be the
simplest model and offers great ease of navigation (Fig. 3).
The ETL phase is the first phase of the realization of our decision-making system,
its purpose is to build the data warehouse. In this respect, we approached “Pentaho
Data Integration” (PDI) [19] which is a complete tool with advanced features of ETL
processing, it allows to create processes composed of data import and export, and
different operations such as conversions, joins, applying filters, or even executing
JavaScript functions. PDI has many connectors, both read and write, allowing it to
access a large number of databases and all types of files. It is by means of this tool
that we realized this phase ETL.
We applied, to the data extracted from the access and error log, the necessary
transformations and operations such as sorts and joins before loading them into the
data warehouse. We started by loading, first, the dimension tables before loading the
fact table in order to respect the integrity constraints when loading the foreign keys
in the fact table (Fig. 4).
After building and feeding our data warehouse, comes the OLAP cube development
step. A cube is a multidimensional structure that contains dimensions and measures,
the dimensions define the structure of the cube while the measures provide the numer-
ical values that interest us. This step is essential in any decision-making system since
it allows us, thereafter, to perform the analyzes on the performance indicators.
For this purpose, we used the Workbench Schema [20] that provides a graphical
interface to perform this task.
Anomaly Detection from Log Files … 521
the malicious users who attempt to inject dangerous characters into server variables
(Fig. 7).
On the other hand, this proposal gave us the opportunity to follow the web server
performance indicators, namely the most serious errors, monitoring server error
reasons, identifying malicious client IPs…
Through this analysis, we’ll be able to understand the HTTP errors (by the isolation
of access logs that contain 400- and 500-level status codes), and on what pages they
occurred so we can fix the problem.
6 Conclusion
In this paper, we present a methodology for analyzing large volumes of log files
based on data warehousing techniques. This multidimensional analysis technique
allowed us to collect the events from several log files, saved and exploited them
to extract useful security knowledge, such as attempted attacks (DDOS, SQLI),
and error detection. It makes easier performing some system tasks like monitoring,
diagnosis, and creating reports about the current state of the system. The obtained
results are encouraging, our methodology allows to keep the data collected and
integrated from different log files and stores a large number of records necessary for
long-term analysis, however there are some difficulties concerning the identification
and the determinations of performance indicators and the attacks through the log
data. Our future work will focus on use of machine learning techniques on log data
analysis.
References
1. Oussous A, Benjelloun F-Z, Lahcen AA et al (2018) Big data technologies: a survey. J King
Saud Univ-Comput Inf Sci 30(4):431–448
2. Lee I (2017) Big data: dimensions, evolution, impacts, and challenges. Bus Horiz 60(3):293–
303
3. Zhang D (2018) Big data security and privacy protection. In: 8th international conference on
management and computer science (ICMCS 2018). Atlantis Press
4. Chandra S, Ray S, Goswami RT (2017) Big data security: survey on frameworks and algorithms.
In: 2017 IEEE 7th international advance computing conference (IACC). IEEE, pp 48–54
5. Mavridis I, Karatza H (2017) Performance evaluation of cloud-based log file analysis with
apache hadoop and apache spark. J Syst Softw 125:133–151
6. Kandpal N, Bandil DK, Shekhawat MS (2020) Analysis of web server logs of websites using
different web mining tools. Solid State Technol 1492–1506
7. Rao RS, Arora J (2017) A survey on methods used in web usage mining. Int Res J Eng Technol
(IRJET) 4(5):2627–2631
8. Amar M, Lemoudden M, El Ouahidi B (2016) Log file’s centralization to improve cloud
security. In: 2nd International conference on cloud computing technologies and applications
(CloudTech). IEEE, pp 178–183
9. Moh M, Pininti S, Doddapaneni S, et al (2016) Detecting web attacks using multi-stage log
analysis. In: IEEE 6th international conference on advanced computing (IACC). IEEE, pp
733–738
10. Salama SE, Marie MI, El Fangary, LM et al (2011) Web server logs preprocessing for web
intrusion detection. Comput Inf Sci 4(4):123–133
11. Azizi Y, Azizi M, Elboukhari M (2018) Tracking attacks data through log files using MapRe-
duce. In: International conference Europe Middle East & North Africa information systems
and technologies to support learning. Springer, Cham, pp 331–336
12. Azizi Y, Azizi M, Elboukhari M (2019) Log files analysis using MapReduce to improve security.
Procedia Comput Sci 148:37–44
13. Nagdive AS, Tugnayat RM, Regulwar GB et al (2019) Web server log analysis for unstructured
data using apache flume and pig
14. Nagaraj K, Killian C, Neville J (2012) Structured comparative analysis of systems logs to
diagnose performance problems. In: Proceedings of the 9th USENIX conference on networked
systems design and implementation. USENIX Association, p 26
524 Y. Azizi et al.
15. Cho M-H, Yoo JS (2017) Exploring online students’ self-regulated learning with self-reported
surveys and log files: a data mining approach. Interact Learn Environ 25(8):970–982
16. Gad I, Manjunatha BR et al (2017) Hybrid data warehouse model for climate big data analysis.
In: 2017 international conference on circuit, power and computing technologies (ICCPCT).
IEEE, pp 1–9
17. Hu Y, Gunapati VY, Zhao P et al (2016) A nonrelational data warehouse for the analysis of
field and laboratory data from multiple heterogeneous photovoltaic test sites. IEEE J Photovolt
7(1):230–236
18. Souza LT, Girish UR (2015) Error log analytics using big data and MapReduce. System 3:4
19. PDI Homepage. https://community.hitachivantara.com/docs/DOC-1009855-data-integration-
kettle. Accessed 15 June 2019
20. Workbench Homepage. https://help.pentaho.com/Documentation/8.2/Products/Schema_Wor
kbench. Accessed 15 June 2019
21. Pentaho BI server Homepage. https://help.pentaho.com/Documentation/7.0/0H0. Accessed 15
June 2019
Web-Based Techniques and E-Learning
A Review on Content Based
Recommender Systems in Tourism
Abstract Recommender systems (RSs) have been used worldwide in several fields
to facilitate the tourists’ planning activities. Tourism is one of the fields that uses RSs
to reduce the overload of information to end users. Accordingly, tourists get recom-
mendations that are most suitable to their profiles. This paper presents a detailed
overview of the tourism recommender systems that were developed since 2008. It
focuses mainly on the content based systems and their applications in the tourism
field.
1 Introduction
Due to the huge amount of information presented online, end users find the task
of planning their touristic activities very tedious. It is because they might not have
the required knowledge or time to go over all search results and select the more
suitable ones for them. RSs have come up with the best solution to this issue to help
end users deal with the information overload. These systems provide end users with
recommendations that suit their registered profiles and are aligned with the behavior
of the rest of the registered community [1].
All over the literature, during the past decades, significant advances in recommen-
dation technology have been achieved. Therefore, RSs have become widely used in
several fields like e-commerce. Tourism is one of these fields where RSs offer services
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 527
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_48
528 A. Bentaleb et al.
to end users like lists of points of interests, trip advisories and tourists’ packages that
much their profiles and preferences [2].
As far as the added value of this paper is concerned, it has been written for the
following purposes: Since in the literature there is not a state of art that is recent and
handles the methods we will need in our contribution, we have accomplished this
survey that responds to our needs. Furthermore, our work will facilitate the work
of other researchers, instead of redoing the same work of examining the literature;
they can start directly by working on their contribution. Also, this survey compares
articles that have proposed contributions to the field of tourism recommendation
especially in Sect. 4.
This survey is organized as follows: it begins by introducing recommender systems
in general, along with their architecture and most known approaches, later it describes
in a detailed manner the content-based approach for recommendation. Then, in the
related works section, it cites some surveys and determines the differences and
common points with them. This is to position the current survey among the already
existing ones found in the literature and that have discussed a similar topic. Finally,
in the applications’ section, it presents some of the well known RSs tourism applica-
tions in the literature that used the content-based approach merged with some other
techniques.
2 Background
Most papers that were published in that sense have stated three approaches of Recom-
mendation: Content-based Filtering, Collaborative Filtering, and Hybrid Filtering as
shown in Fig. 1.
Content Based Approaches Found in the Literature
Definition: Content-based recommendation systems suggests items to end users
based on their interests represented in their profiles and the items’ description.
Users can build their profiles through filling forms, but mainly, this is done through
collecting feedback of these users on items. In the next section, a general architecture
of content-based systems will be presented along with its main components.
The General Architecture of the content based recommender systems: Content- based
recommender systems encapsulate adequate methods and techniques to ensure good
representations of items, generate user profile and consequently compare this latter
with the items. Figure 2 illustrates this architecture. According to this architecture,
the process of recommendation follows these three steps:
– Content Analyzer: the main objective of this step is to generate adequate repre-
sentation of the content items to be usable for the next step in the process. This
pre-processing step is required because sometimes items data has no structure.
The result of this step is the input of the profile learner and filtering component.
– Profile Learner: this component gathers elements representing the users’ prefer-
ences and generalizes them to build the user’ profile. This is done thanks to the
machine learning techniques [11] that can infer the user model based on the items
they like or dislike in previous interactions with the system.
– Filtering Component: this module suggests adequate items to users through
matching the user’s profile with the items’ representation. This is performed using
the similarity metrics [12]. For instance, the matching can be done using the cosine
similarity between the vectors of the prototype and items.
Items can be found in two main representations: either structured or unstructured
[13]. In the first case, items are represented in a database table where each record
stores values for attributes that have a known set of values. Also, according to [10]],
the representation of items can be called structured when they are labelled by the
same set of attributes that can take a known set of values. In such situation, several
ML Algorithms can help in deducing the profile of the user [10]. In the second case,
the descriptions of items are in the form of text deduced from emails, websites,
articles or descriptions of products [10]. Unstructured data does not have attributes
with defined values. Also, the fact that a word can have synonyms and be polysemous
(having different meanings depending on the context) might increase the complexity
of natural language and create enormous issues when learning the profile of a user.
Hence, the old method which is the keyword based profile capturing is not fully able
to deduce the end users’ preferences as they are first managed by operation of string
matching [10].
To overcome these limitations, many methods were suggested throughout the liter-
ature: knowledge bases (ontologies or lexicons) in order to provide a representation
of items and profiles to produce a semantic interpretation for users’ needs.
A Review on Content Based Recommender Systems in Tourism 531
Hybrid Approaches
Definition: The hybrid approach combines the collaborative filtering and content
based method in order to produce recommendations with higher accuracy. This
method has overcome the shortcomings of the collaborative filtering and content
based filtering when used individually. Nevertheless, the task of combining these
two methods in one recommendation process has turned out to be complex and time
consuming [9].
The mixture of these methods can be done in the following manner: Implementing
the algorithms separately and combining the outcome; Using the method of content
A Review on Content Based Recommender Systems in Tourism 533
based filtering in combination with the collaborative one; Using the method of collab-
orative based filtering in combination with the content based one; Building a unique
frameworks that uses both approaches [6].
Techniques of Hybridization
Weighted hybridization: This method generates recommendations through
combining results from several recommenders. It merges the scores of each of the
used techniques through a linear formula. P-tango is an example of the weighted
hybridized method. This system is made of content based and collaborative recom-
menders. Initially, they are assigned equal weights, that later are adjusted as
predictions are confirmed or not. The advantages of the weighted hybridization
method is that all strengths of the used recommender system are used during the
recommendation process in a smooth way [18].
Switching hybridization: According to this approach, a heuristic algorithm is
applied to swap to the most suitable recommendation techniques to generate a better
rating. This technique has been proposed to overcome the shortcomings specific
to one recommendation method such as the “new user problem” of content-based
recommendation, in this case, it switches to a collaborative recommendation system.
The main benefit of this switching hybrid method is its full awareness and sensitivity
to the strengths and weaknesses of its recommending components. However, the
main disadvantage of this approach is that it adds more complexity to the recom-
mendation process by integrating the switching criterion among the parameters of
the recommendation system [19].
Cascade hybridization: Within this approach, the recommended items are ordered
and refined through an iterative process. The recommendation list output by a tech-
nique are refined by another recommendation approach. Usually, the first list is
coarse. This hybridization technique is efficient and acceptant to the noise created
by the iteration process [20].
Mixed hybridization: This method combines recommendations generated from
several techniques instead of assigning one recommendation per item. Recommen-
dations associated to items are generated from different recommendation techniques.
Within this approach, individual performances do not affect the general picture [19].
Feature combination: The features generated by a certain recommendation tech-
nique are given as inputs to another recommendation technique. For example, in
collaborative filtering, the feature that is the ratings of users’ similarity is consid-
ered, in a case-based reasoning recommendation technique, among the features used
to generate the similarity between items [6].
Feature augmentation: The inputs of this method are the ratings and informa-
tion resulted from the previous recommender. The feature augmentation hybrids are
superior to the methods of feature combination as they add a few number of features
to the initial recommender.
Meta level the internal model: This method is generated by a recommendation
technique and used as input for another one. The generated model is richer in data
than single ratings. This type of hybrids had solved the issue of sparsity faced with
534 A. Bentaleb et al.
collaborative filtering techniques. This is because it uses the entire model learned by
the previous recommendation technique as input for next one [6].
3 Related Works
Throughout the literature, several authors have published surveys concerning the
systems of recommendation in the tourism field.
“Personalizing recommendations for tourists” is a survey proposing a state of
art of the tourism recommender systems (until early 2009). The presented systems
are grouped under several criteria including: the used personalization approaches
including mainly (Bayensian networks method and tools of decision making); the
chosen approach of recommendation (content, collaborative or hybrid approach); the
hardware devices, or the items used in for recommendations (restaurants, flights or
hotels). As a matter of fact, this survey presents these various systems per categories
but in a general way. It did not provide a detailed technical explanation on how they
have been built, therefore, it did not relate to the latest advances of technology lately
(exploration of social networks, advanced tools of geolocalization.) [2].
This survey presents an examination of the tourism recommender systems from
the perspective of sociologies, not from the innovative point of view. This paper
claims that smart frameworks are fundamental in the tourism industry as it includes
several aspects to be dealt with: the fact that data sources are distributed, the new
environments are full of risk and uncertainty; the tourists are mobile most of the
time. Within this survey, the author presents some systems that have dealt with these
issues. It also, provided an evaluation as far as the user interaction, the social side,
the context, and decision making mechanisms are concerned, to provide the best
services to end users. Nevertheless, this overview did not tackle the technical side of
those systems [21].
“The planning of cycle trips in the province of East Flanders” is another review
that presents some frameworks that have been introduced from the years 2001 to
2011 serving at building plans of trips. This survey’s authors compared the cited
systems regarding their various functionalities that are mainly planning trips. They
have used the following criteria for that concern: budget, weather circumstances, and
mandatory points of interests, hotel selections and routing. This paper tackles the
orienteering issue and how its extensions can be utilized to demonstrate trip planning
functionalities [22].
“Intelligent tourism recommender systems: A survey “ is another overview of the
tourism recommender frameworks. This paper has presented the most known smart
e-tourism recommender systems that have been published in the conferences and
Artificial Intelligence journals since 2008. Within this survey, the authors considered
different criteria including: the functionalities provided by these systems, the used
artificial intelligence approaches, the various recommendations algorithms used, and
different kinds of interfaces. As far as the details of building these recommendation
systems is concerned, this survey has presented some guidelines in that sense, it also
A Review on Content Based Recommender Systems in Tourism 535
mentioned that there might be some interesting future works in that fields for the
coming years [23].
Here is another survey that has presented a review of the state of art in that
field especially the mobile tourism recommendation systems. It provided insights
concerning the services they offer, the challenges they encounter along with the
promising areas of further research that can be conducted in the fields of mobile
recommender systems in tourism [24].
“A Short Survey of Recommendation Technologies in Travel and Tourism” is
another survey found in the literature tackling the same topic as this. It starts by
presenting the different paradigms of recommender systems used in e-commerce
applications. There are three approaches: the collaborative filtering, content based
and the hybrid paradigm, along with their advantages and disadvantages. Applica-
tions and examples have been used to better explain the paradigms. Then, the authors
moved to discussing the different applications of recommender systems, especially in
the tourism field. They have chosen that field since it is very important and dynamic,
and lately, it contributes to more than 50% of the global B2C turnover. They discussed
the shortcomings of existing applications. Compared to our survey, this one lacks the
detailed aspects. This has been noticed in the background part while presenting the
paradigms, it gave a short overview of them, also, when discussing the applications
of tourism, they used few ones, while there exist so many systems in the literature
that have implemented the recommender systems [25].
“A survey on mobile tourism Recommender Systems” is another survey found
in the literature and has tackled a similar topic, which is tourism recommenda-
tion mobile systems. It has presented a review of state of art of these systems,
and suggested a classification of mobile tourism recommendation systems. Later,
it discussed the possible research directions and the encountered challenges when
dealing with tourism mobile recommender systems. Notably, the accuracy of context
aware mobile RSs, the measurements of the quality of recommendations and consid-
eration of new aspects when planning tours/ routes for users. Compared to our survey,
this one focuses on mobile applications mainly [26].
“Recommender System Application Developments: A Survey” claims the impor-
tance of recommender systems as they reduce the overload caused by selecting the
most relevant services to end users from a tremendous amount of data based on their
preferences and profiles. It presents a review of the existing applications of recom-
mender systems in several categories: e-tourism, e-government, e-business. also, it
tackled the main approaches and techniques used in recommender systems. At last,
the authors identified new topics and directions for researchers to take. For the arti-
cles, this survey used two types: the first type is articles dedicated to explaining
the techniques and approaches used in recommendations systems while the second
type presented the different applications and case studies concerning recommender
system applications. The filtering and selection process of these articles have been
fulfilled in four steps. Compared to our survey, this one has tackled the topic of
recommender systems from a broad aspect as it included applications from several
domains. While our survey focuses in the tourism fields, this one handled the topic
of RSs in a general manner [27].
536 A. Bentaleb et al.
Before going on a trip, end users prepare their plans before travelling. They search
for details of possible destinations, transportations, hotels, book reservation. As it
seems, it is very tedious to examine, estimate and compare all these aspects of
planning a tour. Here comes, the need of assistance to overcome the complexity of
this time consuming task; there have been various tourism recommender systems in
the literature that facilitate the operation of planning such activities [29].
Here are some of the most known applications of recommender systems in the domain
of tourism and more specifically in the activity of planning:
“A hybrid recommendation approach for a tourism system” has presented a recom-
mender system for tourism using a methodology where classification based on associ-
ation is used. It is called Associative Classification Methods; it is about using concepts
from both association and classification to be able to employ association rules in a
prediction context. The associative classification could provide more reliable recom-
mendations on sparse data contexts whereas fuzzy rules allowed classifying a user
into multiple groups. Throughout case studies, it has been proven that such method-
ology could enhance the quality of generated recommendations and overcome the
shortcomings of already existing recommendations approaches [30].
“Sightseeing planning system (PSiS)” is another system that has been developed
to help users in planning what should be done in a certain place. It is based on several
aspects such: the location and the purpose of the visit, the time details and the weather
forecast. This recommender system is useful as it provides end users with refined
suggestions while they are planning for a tour or in real time when they are actually
on a trip [31].
“eCompass” is another application, that has been designed, to help users through
the public transit. It has been presented in both mobile and web platforms. It is a
service oriented architecture approach where outputs are given by end users in order
A Review on Content Based Recommender Systems in Tourism 537
to suggest a tour. A service for multimodal planning has been used in addition to
yahoo weather forecast services in order to plan trips. The resulted recommendations
for the tour is presented in either a map or a list view [32].
“Filter first; tour second” is another framework that has been proposed to produce
adequate recommendations. Its processing is based on online data sources and infor-
mation from social media. For the points of interest, there are some predefined ones
based on users’ preferences, while the rest is identified through the collaborative
filtering approach. This is done to maximize the satisfaction of end users. This
framework is based on two modules: 1) collaborative filtering engine to produce
the mandatory points of interests along with the optional ones, based on the popu-
larity of this POIs on social media and online reviews. 2) The iterated Tabu search
is used to generate the daily tours [33].
Some of the applications that offer travel packages recommendations are: “Per-
sonalTour” is a multi agent recommender system that has been built to produce
adequate travel packages based on users’ preferences. This framework uses the AI
paradigm. The knowledge bases of this system store the recommendations of the
users along with their reviews of the offered travel services. The PersonalTour agent
transforms these reviews to confidence degrees given to the various services (flight,
hotel, attraction services) [34].
Another travel package recommendation framework is the “Object oriented RS”,
which generates personalized travel packages. This framework has proposed a new
way to represent travel packages where the object is the collection of feature value
pairs. In order to mine the implicit similarities between objects, two models are used:
The Objected Oriented Topic Model (OTM) and the Objected Oriented Bayesian
Network (OBN). Later on, the recommendations are generated based on these models
and using the collaborative filtering approach [35].
This application provides recommendations of personalized travel package for
several POIs using the crowd sourced user footprints gathered from the Jie Pang
system. The latter platform is used very widely in china to “check in” and sharing
locations, events and feedback about services. This application offers travel packages
with various types of locations with different sequences. The model of the data (user
preferences and POIs) is done through the spatiotemporal features used in check-in
records. Therefore, recommendations are generated based on user’s demands and by
joining the users’ preferences along with the spatiotemporal constraints [36].
538 A. Bentaleb et al.
In case of having a group of end users who would like to get recommendations before
starting a touristic activity, the following applications have brought some pertinent
solutions to help them in that:
“e-Tourism” is a web based application to provide tourists with recommenda-
tions concerning tours. This platform has been implemented specifically for the city
of Valencia and can be used by a single or a group of users. It is designed as a
set of modules. The main one is the “Generalist Recommender System Kernel”
(GRSK) which is responsible of generating recommendations in case of single user.
The system uses a hybrid approach that combines (demographic, content-based and
general likes filtering) to generate recommendations based on users’ preferences. In
the case of a group of users, recommendations are computed by the application of
the intersection and aggregation mechanisms on the individual preferences [37].
“A group recommender system for tourist activities” is another application that
has presented a promising method for generating adequate recommendations for a
group of users. During the process of generating recommendations, it combines the
following elements: group tastes, the users demographic classification and places
they visited in previous trips. The generated items are computed from the indi-
vidual personal recommendations. This platform combines the use of the following
technique: intersection, aggregation, and incremental intersection [38].
“A multi agent architecture for single user and group recommendation in the
tourism domain”, this paper presents a Multi Agent System used for planning touristic
activities for single or group of users at a given region. The agents interact to learn
dynamically the users’ profiles to generate a list of proposed activities for the end
users or groups based on the interaction of users with similar ones and the system
[39].
“Social group recommendation in the tourism domain”, this work suggests
a promising approach for generating recommendations as it includes not only
users/groups preferences, but considers the social relationship among them. Profiles
are built based on a combined approach that includes the demographic information
of users, the feedback of the community (collaborative filtering) and the descriptions
of the items to be recommended (the content based filtering). This hybrid profile
is used to generate ratings of unknown items. In the case of groups, the individual
ratings are weighted based on the social relationships between the group members
and the preferences of the rest of the team. Then, the estimations of the group are
obtained by combining the ones of the group members [40].
To help end users determine the most interesting points of interests, the following
applications have been designed for that purpose:
A Review on Content Based Recommender Systems in Tourism 539
5 Conclusion
References
Abstract With the rapid evolution of SOA (Service Oriented Architecture) and
wide techniques of web services, the web service semantic has attracted the most
attention during recent years. The WS (Web Service) has become a promising tech-
nology for the development, deployment, and integration of Internet applications. The
large-scale deployment of web services brings great value to web service customers;
however, the large-scale, dynamic, and heterogeneous web services make it difficult
to discover and compose WS. Therefore, as one of the challenges to interpreting rela-
tionships of semantics is to understand the functional associations between ontology
concepts. Several approaches have been described to extract relationships between
ontology concepts using term-matching methods. However, it is necessary to have
flexible and efficient methods to ensure functional relationships. To facilitate the
discovery the relevance of a service to a query, there are many solutions based on the
use of Semantic Web technologies to develop homogeneous service descriptions to
reason on and to support precise and flexible discovery. In this paper, we proposed
an approach to measure semantic similarity by combining Latent Semantic Analysis
Similarity (LSA) and IO-MATCHING semantic matching. The density clustering
services method is treated as a pre-processing method for semantic matching between
the web service and the user requirements. The results of the simulation show a high
accuracy to cluster semantically web services.
1 Introduction
With the growing utilization of web services, the research community has attracted
a lot of attention in the area of service discovery and composition [1, 2]. However, it
remains difficult to accurately obtain the necessary information and knowledge from
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 543
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_49
544 N. El Allali et al.
the data. WS registry search provides limited results for conventional approaches for
WS discovery due to the general unavailability of semantic descriptions of services.
Besides, with a growing number of Web services, these approaches generally result
in low productivity with marginal performance. For that reason, the majority of
the community is oriented to Semantic Web technologies (Fig. 1) [3], for example
RDF (Resource Description Framework) [4], RDFS (Resource Description Frame-
work Schemas) [5], OWL (Ontology Web Language) [6], to enhance Web services
as semantic descriptions. The combination of Web services and Semantic Web
technologies leads to the Semantic Web Services (SWC) concept.
Embedding Web Services in the Semantic Web domain is becoming ever more
interesting [7]. Using the concepts of the Semantic Web in the Web Services domain
provides a tool for using the meaning and significance of the data exchanged to
improve the feasibility of different tasks such as discovery, composition, etc. SWC
can be defined as Web services with semantic descriptions. This semantic is provided
through the ontologies, which is one of the most important technologies of the
Semantic [8]. It aims to describe concepts and the relationships that link them to
each other, and with rules of deduction to make them more understandable and
relevant.
The Latest solutions [9, 10] that aim at minimizing the discovery duration and
enhancing the precision ratio are becoming more important by clustering to reduce
the large-scale of the web services and to facilitate service discovery. Although
clustering is one of the most successful approaches of data mining [11], analysts
neglect sufficient resources for rapid, iterative clustering analysis, which is important
for the creation of conclusions and data rationale. A precondition for clustering
algorithms is the correct definition of a set of measures which exactly evaluate the
semantic similarity between services. under these conditions, Web service similarity
measurement and clustering has become a challenging issue. Therefore, to overcome
the problem of inaccurate service discovery in a fast and efficient way. We define a
combination of semantic similarity by subsumption reasoning and LSA, taking into
consideration the service clustering based on the DBSCAN algorithm [12] to reduce
the time of semantic matching between the requested service and massive service.
The remainder of this paper is organized as follows: The next section gives an
overview of related work. Section 3 introduces our proposed approach. Section 4
presents the experimental results. Section 5 points out the conclusion with the future
works.
2 Related Works
This section presents the different works that aimed at minimizing the discovery
duration and improvement the accuracy ratio for having an efficient system, with the
techniques of data-mining, in particular clustering analysis [13], can be used to group
similar services, it allows to reduce the matching time between services and to accu-
rately classify the web services in the repository. A large share of the studies focused
Towards Semantic Web Services Density Clustering Technique 545
Fig. 1
XML-S
Standard technologies for OWL
semantic web services
RDF-S
XML
RDF
HTTP URI
on this area develop new ontological methods using different Semantic similarity
algorithms [14, 15]. This paper is closely related to the semantic similarity measure
with the discovery service method by clustering. For that reason, we only present
related works in this direction:
In [16], the authors presented a new web services discovery mechanism based
on semantic matching, and the clustering of services is brought forward for efficient
and appropriate web services discovery, which combines functional similarity with
process similarity. Their proposed approach relies on the information available from
the semantic description model, based on the improvement of Lin’s [17] semantic
similarity measure to include opposition or degree of contrast as defined in [18].
Their objective is to develop the efficient WS discovery method by pre-clustering
that allows them to perform semantic and scalable WS discovery in a minimum of
time and thus reduce the search space. According to their results are acceptable in
terms of time compared to the results obtained by [19], but it remains a difficult
task to depend on semantic dictionary such as WordNet to obtain the oppositeness
relationships between concepts, which allows for not finding adequate services by
simple methods, and thus the algorithm is not accurate to describe semantically due
to the ignorance of the antisense relationships between two concepts.
In the work of [20] focuses on the web services clustering before discovering a WS
based on the OWL-S description language. They use two aspects to measure service
similarity (functional and process) for the aim of clustering services. A weighted
domain ontology approach has been used to compute functional similarity by using
I/O parameters with domain ontology developed with the help of WordNet and
available web ontologies. According to their experimental results showed for the
clustering algorithm based on the matching method of web service I/O [21] as an
efficient way to discover a WS.
The authors [22] proposed a hierarchical clustering approach for semantic WS
discovery. The main objective is to enhance the efficiency and accuracy of traditional
Web services discovery with the use of the vector space model. The Web services
are converted into a standard vector format via WS descriptions. Then, a semantic
analysis is used to reduce the size of the vector term with the help of WordNet. It
has performed the semantic expansion to respond to the user service need.
The authors [23] proposed a WordNet-VSM model to measure feature similarity
by generating service name, process and message vectors. Using unsupervised neural
546 N. El Allali et al.
networks based on kernel cosine similarity estimation rather than conventional clus-
tering algorithms. However, this approach has some challenges with clustering, such
as the output node linking to the weight and requiring a lot of processing.
In [24] proposed a service clustering approach inspired by the actions of ants.
Using a semantic service description related to a domain ontology that includes
concepts was used to annotate service input and output parameters. A set of metrics
was defined to compute the degree of match between services [21]. This approach has
been checked and validated on all the SWC available in the SAWSDL-TC benchmark
collection. The experimental results showed that there are many parameters that affect
the behavior of the service clustering algorithm.
As presented in [25], a new WS discovery approach is proposed, where common
topic groups are extracted from the service topic distribution matrix created by
topic modelling methods, and these topic groups support the process of matching
user requests with appropriate web services. The proposed approach can preserve
service discovery performance at a high level by considerably reducing the number
of candidate Web Services, consequently reducing response time.
The work proposed by [26] is to measure similarity based on WordNet in order to
determine the similarity between the service name and the text description using input
and output parameters based on the ontology of the domain. They use the OWL-S
files to extract the semantic information and the service similarity computed through
it. They are used as a K-medoids clustering algorithm.
In [27] proposed a method to facilitate cluster discovery of web services using
WSDL document meta tags. The efficiency of clustering was maintained by inco-
herent distribution and classification of noise tags. Moreover, the authors proposed
automated clustering of web services by tagging web services on a domain search
engine or UDDI or ontology. However, syntax-based approaches that have a lower
performance disadvantage due to the complexity of natural language, a variety of
semantic approaches are also suggested.
Calculate the distance between the q and each cluster head, select the closest cluster
similarity to obtain the search results. At the third level, DBSCAN pre-clustering of
services to limit the search space and thus enable rapid semantic matching in large-
scale service pools. Generally, the performance of Web services discovery could be
improved by semantic matching and the rate of discovery could be improved by
pre-clustering.
Semantic similarity measures of Web services offer many advantages and the majority
of the WS discovery methods depend on measuring the similarity of the wanted
service with each of the advertised services. The community of Semantic Web
Services has evolved many models for semantically representing global web services
such as Ontology Web Language for Services (OWL-S) and WS Modelling Ontology
(WSMO). However, it remains a challenge to evaluate the similarity between two
concepts in order to make an accurate selection of the most appropriate service. The
efficiency of clustering relies on the way similarity measurement has been performed
between data. In addition, there are solutions [16–26] based on semantic dictionaries
such as WordNet to interpret the relationships existing in the ontology between
two concepts, but it’ is limited to precision and automatization to describe services
semantically. With the description language OWL-S provides more flexibility and
efficiency semantic matching compared to WordNet. For another part, most of the
measures based on distance are based on counting the number of arcs that separate
two directions in a taxonomy such as the work of [28, 29], which just depend on
548 N. El Allali et al.
Definition 2 (Web Service): Web Services are described by the OWL-S ontology. It
may be realized as a 4-tuple: W S = S N ame |ST ext Des |I n S |Out S Where S N ame is the
service name, I n S and Out S are the input and output concepts respectively, ST ext Des
is a summary of what the service is providing.
Definition 3 (Latent Semantic Analysis (LSA)): the LSA [30] is based on a multi-
dimensional representation of the meaning of words in language. The idea behind
it is that words in a text corpus appearing on the same linguistic contexts have
similar meanings. It uses the algebra technique The Singular Value Decomposition
(SVD) which deals the transformation and decomposition of a matrix. Formally, the
word-service matrix decomposition of matrix A can be defined as:
A=U VT (1)
Where,
A is Input matrix (m × n)
is Words × extracted concepts (n × n)
U
Scaling values, diagonal descending matrix (n × n)
V T Documents × extracted concepts (n × n)
Exploiting OWL-S textual elements such as “<profile:textDescription>” and
“<profile:serviceName>” to provide us with more information for semantic
matching. To evaluate the similarity between two services, with a massive volume
of data, using Latent Semantic Analysis to reduce the dimensionality of our service
vectors instead of directly comparing the cosine similarity of the bag-of-word vectors.
After having a decomposed matrix, requires to use cosine similarity between services
either at the text description name or service name level. The measurement of cosine
similarity is defined as follows:
Si · S j
Sim Si , S j = (2)
Si S j
j j
Sim T extual Si , S j = ρ × Sim Descri ption STi ext Des , ST ext Des + σ Sim N ame S iN ame , S N ame (ρ, σ ) ∈ [0, 1]
(3)
(4)
where C1 and C2 are the services concepts respectively. to calculate the average
similarity of service concepts pair I/O is done as follows:
m
j
i
i
Match SC1 , SC2
Sim Match Si , S j = (5)
m
where m is number of concepts.
Fun_sim Si , S j = α × Sim T extual Si , S j + β × Sim Match Si , S j (α, β) ∈ [0, 1]
(6)
where α is the textual similarity weight and β is the Semantic Similarity Degre
weight.
database, the DBSCAN algorithm allows us to detect any form in the search space,
unlike other clustering algorithms based on distance. The fundamental concept of
the DBSCAN is to identify high-density regions which are separated from each other
by low-density regions. The main parameters of the algorithm are eps (Maximum
radius of the neighborhood) and the MinPts (minimum number of points). The
concept of the eps neighborhood of Sj is given by:
Neps S j = S j ∈ D|dist Si , S j ≤ eps} (7)
To calculate the service similarity is done by the distance d i st S i , S j between
two services S i and S j :
dist Si , S j = 1 − Fun_sim Si , S j (8)
A point can be considered as a core point if the number of samples in the eps-
neighborhood is more than or equal to the MinPts density threshold. When there is
a core point in the neighborhood of a non-core point, the sample is referred to as a
boundary point. Noise is a data point that is neither a core nor a boundary point. The
core point set is referred to as d c is expressed by:
dc = S j ∈ D|Neps S j ≥ Min Pts} (9)
The noise is the set of items in D that make no relation to any Ci cluster, this set
denoted as dn :
/ Ci }, i = 1, 2, . . . , l
dn = {Sk ∈ D\dc | ∀i : Sk ∈ (11)
We used the OWLS-TC version 4.0 dataset [31] to evaluate our proposed approach.
The dataset includes of 1083 indexed OWL-S services, that have all been collected
from the public IBM UDDI databases and semi-automatically converted from WSDL
to OWL-S. The experiments have been tested on a PC with an Intel ® Core (TM)
i7-4770 CPU and 8 GB main memory running Windows 10 operating system. Use
Java language programming to implement the DBSCAN Semantic Web Service
Clustering. The Parameters of the proposed method is defined as follows in Table 1:
To evaluate the performance of the proposed method, we have compared our
results with other work of [16].These results have been analyzed in terms of preci-
sion, recall, and F-Measure to prove the effectiveness of clustering approaches with
a suitable semantic similarity measurement method.
The results in the Fig. 3 shows that the proposed similarity method is highly
performant in terms of precision when the threshold of “eps-neighborhood” is equal
to 0.45, that means the performance of clustering web services will be more efficient
and more similar when the precision value is equal to 84.25% and the recall is equal
to 82.15%.
From the analysis of the results, it is possible to conclude the similarity of the web
services in this article is more efficient to support in either the discovery process or
in the composition phase. As compared to other approaches relying on the semantic
dictionary like WordNet as a solution but with less accuracy to find web services as
mentioned in Sect. 3.
0.8 Precision
Recall
0.6 F-Score
0.4
0.2
0 eps
0.25 0.35 0.45 0.55 0.65
5 Conclusion
In this work, we proposed a similarity method to group semantic web services effi-
ciently and accurately, as it enables us to quickly find web services either in the
discovery or in the composition process. Experimental results were more encour-
aging in terms of precision, recall, and f-score. As a future step in our work, it
consists of evaluating the effectiveness of this proposed approach during the compo-
sition process, since discovery is a more important phase to quickly find relevant
services, on the other side focusses mainly on enriching our model with semantic
descriptions by other functional aspects such as pre-conditions and post-conditions
and thus making the matching scheme more intelligent.
References
18. Mohammed SM, Dorr BJ, Hirst G, Turney PD (2011) Measuring degrees of semantic
opposition. Technical report
19. Kokash N (2006) A comparison of web service interface similarity measures. In: STAIRS, pp
220–231
20. Xie L, Chen F, Kou J (2011) Ontology-based semantic web services clustering. In: 2011
IEEE 18th international conference on industrial engineering and engineering management,
pp 2075–2079
21. Paolucci M, Kawamura T, Payne TR, Sycara K (2002) Semantic matching of web services
capabilities. In: International Semantic Web Conference, pp 333–347
22. Gao H, Wang S, Sun L, Nian F (2014) Hierarchical clustering based web service discovery. In:
International conference on informatics and semiotics in organisations, pp 281–291
23. Chen L, Yang G, Zhang Y, Chen Z (2010) Web services clustering using SOM based on kernel
cosine similarity measure. In: The 2nd international conference on information science and
engineering, pp 846–850
24. Chifu VR, Pop CB, Salomie I, Dinsoreanu M, Acretoaie V, David T (2010) An ant-inspired
approach for semantic web service clustering. In: 9th RoEduNet IEEE international conference,
pp 145–150
25. Wang J, Gao P, Ma Y, He K, Hung PCK (2017) A web service discovery approach based on
common topic groups extraction. IEEE Access 5:10193–10208
26. Wen T, Sheng G, Li Y, Guo Q (2011) Research on web service discovery with semantics
and clustering. In: 2011 6th IEEE joint international information technology and artificial
intelligence conference, vol 1, pp 62–67
27. Wu J, Chen L, Zheng Z, Lyu MR, Wu Z (2014) Clustering web services to facilitate service
discovery. Knowl Inf Syst 38(1):207–229
28. Rada R, Mili H, Bicknell E, Blettner M (1989) Development and application of a metric on
semantic nets. IEEE Trans Syst Man Cybern 19(1):17–30
29. Wu Z, Palmer M (1994) Verb semantics and lexical selection, arXiv preprint
30. Dumais ST (2004) Latent semantic analysis. Annu Rev Inf Sci Technol 38(1):188–230
31. Klusch M, Khalid MA, Kapahnke P, Fries B, Vasileski M (2010) OWLS-TC OWL-S service
retrieval test collection
KSTAR: A Knowledge Based Approach
for Socially Relevant Term Aggregation
for Web Page Recommendation
1 Introduction
Social tagging is a technique to identify and retrieve the contents available on the
internet based on their common attributes. In this process, a keyword is associ-
ated with any content on the internet, such as website content, social bookmarks,
photos, videos, blogs, etc. It is widely used in social networks, media service appli-
cations, and traditional websites. The users create tags, and there are not any rules
in naming them. Tags are named based on the features or characteristics of the web
content according to the user’s perspective. This paper introduces a novel approach
to generate recommendation tags based on the Random Forest Classification and Ant
Colony Optimization strategies.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 555
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_50
556 D. Surya et al.
Motivation: In this paper, the Web usage Mining technique is employed to extract
users’ profile data. Web usage Mining, a subset of Data Mining, extracts data from
the user’s historical data content on the internet, such as search logs, transaction logs,
social network activity, etc.
2 Related Work
De Caro et al. [1] have put forth a highly scalable tag-based recommender system that
can simultaneously suggest mixed types of content. It also provides suggestions to
enhance Software’s performance and reusability. Dorigo et al. [2] review the develop-
ments and the recent trends in the Ant Colony Optimization research. Dhandi et al. [3]
provide an overview of web mining techniques required to maintain and customize a
web application. Anitha et al. [4] a method to predict users’ behavior by employing
web usage mining techniques on users’ web server log files. It presents an overview
of the different algorithms and methods used to analyze the behavior of the user.
Wei et al. [5] suggested two deep learning-based models that can work effectively
by solving the cold start. The content of the features is extracted using a deep neural
network called SADE. The proposed solution enhances the past experience of the
user and the confidence of recommendation systems. Hameed et al. [6] categorized a
collaborative filtering recommendation system and the required criteria for selecting
the data set. It also discusses the measures to evaluate the recommendation system’s
performance, the uses, and the challenges faced by the collaborative filtering recom-
mendation system. Subramaniyaswamy et al. [7] proposed a novel model that can
generate many movie recommendations compared to other models using filtering
and clustering techniques. Jiang et al. [8] present a reliable novel multi-deployable
slope one algorithm developed by merging reliable data and user similarity approach.
The algorithm used in their work has higher accuracy compared to the conventional
slope one algorithm. Fu et al. [9] developed a novel deep learning-based method used
as a Collaborative Filtering recommendation model. The proposed algorithm gains
the semantic information revealing the correlation between user-user and item-item.
The algorithm also learns the low dimensional user and item vectors. Jain et al. [10]
KSTAR: A Knowledge Based Approach … 557
3 Proposed Methodology
Random Forest is supervised learning technique used to sort the input data items
into predefined categories. It creates a decision tree based on different data samples
available in the data source and predicts it. There are so many trees of such is created
that it resembles a forest. A decision tree is a tree-shaped structure in which each
node performs a test on the attribute, and each branch of the tree represents a predic-
tion outcome of decision analysis. It can handle multi-output problems, i.e., both
numerical and categorical data. The final decision is based on merging each tree’s
prediction.
Q
Txky = (1)
k
Lk
Txky (2)
k
Tx y ← (1 − p)Tx y + Txky (3)
k
β
Txαy ∗ ηx y
pxk y = β (4)
α
z∈allowedx Tx y ∗ ηx y
In Eq. (1), Txy, gives the pheromone secreted by the kth ant moving from x to
y, Q is a constant, and Lk is the cost for the ant’s path. Equation (2) gives the total
amount of pheromone found along the path xy, ignoring the vaporization effect.
Equation (3) gives the total amount of pheromone found along the path xy consid-
ering the vaporization effect: p is the pheromone evaporation coefficient. Txy is the
558 D. Surya et al.
amount of pheromone deposited when the ant moves along xy. Equation (4) simulates
the probability of the ant to choose a path. Txy signifies the quantity of deposited
pheromone when the ant moves alongside the edge xy, nxy = 1/dxy (d is the distance
between x and y). The influence of Txy and the influence of (η)xy is controlled by
parameters α and β respectively.
The user interests are captured and represented to predict the user’s intention. The
user social web data is examined to collect the keywords associated with the user. The
collected terms undergo feature population and synonymization. WordNet 3.0 is used
to perform synonymization. The architecture of the proposed KSTAR framework is
depicted in Fig. 1.
The random forest algorithm is employed on the knowledge base to classify the
content. Each decision tree is constructed based on the CART (Classification and
regression tree) algorithm. The entire training dataset is fed as the input to the root
node of the decision tree. The root node performs the first classification and creates
two child nodes. The input data set is divided into two groups, and each group is
rendered as the input for each node. The child nodes further classify their corre-
sponding input data set into two groups. At each level, the input data set is classified
to obtain the purest possible distribution. This recursive division process repeats to
get a data set with zero uncertainty, i.e., only one label type. The random forest has
been chosen for various reasons. As the final prediction is based on all the decision
tree predictions, it reduces the risk of overfitting, thus increasing the model’s perfor-
mance. It reduces the training data time and can run efficiently on large datasets with
high dimensionality.
4 Implementation
This paper presents an approach to suggest tags that are highly relevant to the user’s
intention. The algorithm’s input is the user’s intention, web usage data, and has
access to their social web data. Incipiently, the topics related to the web usage data
are extracted. The terms generated based on user intention from the user’s web usage
data undergoes the feature population. The terms obtained from the previous step
undergoes synonymization using WordNet 3.0. WordNet is a lexical database of
words, mainly English, linked by their semantic relationships.. Synonymization is a
technique to furnish the synonyms related to a word. This technique has been used
to generate semantically related words associated with the topical words extracted
using the web usage data. WordNet is used to determine the words that are semanti-
cally related to one another. This step is known as feature enrichment. The keywords
generated in the feature enrichment step are used to construct a standard knowl-
edge base like Dbpedia. It has been exploited to extract structured information from
560 D. Surya et al.
the information available in Wikipedia. The information from the database can be
modified and retrieved by a query language like SPARQL. SPARQL is an RDF
(Research Description Framework), a framework originally designed to represent
information in the Web, query language (i.e., semantic query language) for databases
The Real-world knowledge base that has been incorporated in this approach is the
DBpedia. The Random Forest algorithm is employed to classify the contents of the
dataset. Finally, Ant Colony Optimization is applied to each classified content. Upon
successful implementation, the algorithm generates tags that capture the user’s inten-
tion efficiently as the output. Algorithm 1 depicts the proposed Tag Recommendation
algorithm of the KSTAR Framework.
The dataset on which the experimentations for this research work was conducted is
the NASA web access log from 01/07/1995 to 06/07/1995 (six days). The dataset
consists of 4,49,911 records upon filtering and preprocessing. The proposed model is
baselined with VOMMRS [23] which employs Fuzzy C-Means Clustering with Vari-
able Order Markov Model for Recommendation. Fuzzy C-Means clustering works
reasonably well with Variable Order Markov’s Model, but the approach does lag
since they used only Web log data without using auxiliary information. There is a
need for extensive Web log information or the users’ historical information, which
makes this method entirely user centric. The fuzzy C-Means algorithm also has some
disadvantages associated with it. The computational time is relatively high, and it is
susceptible to noise (outlier points). The main limitation of Variable Order Markov
mode is that it cannot express the relations or dependencies between the hidden
states. Table 1 depicts the performance evaluation for the proposed KSTAR and its
comparison with other baseline models like GAAHC, IFWIAR, and VOMMRS.
IFWIAR [21] used Fuzzy Weighted Iterative Association Rule-Based Ontology
Model for query recommendation. A static domain ontology model restricted to a
specific domain, namely the Healthcare domain, has been proposed. The model
performs well, but the fuzzy weighted iterative association rule mining can be
improved based on the amalgamation of the recent algorithms. GAAHC [22] uses the
Group Average Agglomerative Hierarchical Clustering. It performs well over a static
dataset but does not when dynamic knowledge has been supplied to the approach.
The approach focuses on the ranking webpages based on time and frequency. The
modified Levenshtein distance has also been incorporated, and the personalization
is the focus of the approach. However, there is a need for the web search to be made
knowledge-centric and user-relevant. Thus, a newer paradigm is needed.
KSTAR employs Random Forest Classification and Ant Colony Optimization. It
performs much better than the baseline approaches because it captures users’ inten-
tions and extracts socially relevant terms from the Social Web APIs. The topic extrac-
tion is initially extracted from the users’ web usage information. The model focuses
on the feature population based on synonymization and the extraction of trending
terms from the Social Web. Apart from this, the DBpedia knowledgebase populates
the real-world knowledge into the model. The terms and knowledge aggregated based
on the user intentions are used to classify the datasets using the Random Forest, which
is quite efficient when the knowledge density is relatively high. However, ACO is
incorporated to optimize the recommendation sets based on the classified content
by prioritizing users’ web usage data. There is a precise balance between the users’
past information and the available real-world knowledge, which makes KSTAR quite
efficient than the baseline models.
Figure 2 depicts the distribution of accuracy in percentage against the number
of recommendations generated. It is evident from the graph that the suggested
approach has better accuracy than any other baseline models. The precision in the
proposed model is more remarkable than GAAHC by 5.53%, IFWIAR by 6.79%, and
VOMMRS by 18.33%. The recall percentage of the proposed model is higher than
GAAHC by 7.02%, IFWIAR by 7.55%, and VOMMRS by 14.91%. The accuracy
percentage in the proposed approach is greater than GAAHC by 6.54%, IFWIAR by
6.54%, and VOMMRS by 16.57%. The F-Measure percentage of KSTAR is more
significant than GAAHC by 6.96%, IFWIAR by 6.55%, and VOMMRS by 17.37%.
The FDR of the proposed technique is lesser than GAAHC by 0.05, IFWIAR by
0.06, and VOMMRS by 0.18. The filtration rate percentage of KSTAR model is
greater than GAAHC by 12%, IFWIAR by 16.95%, and VOMMRS by 18.67%. The
computational time of KSTAR is lesser than GAAHC by 21 ms, IFWIAR by 10 ms,
and VOMMRS by 22 ms.
6 Conclusions
It is very evident from the results that the proposed method, KSTAR operates much
better than the other baseline models. The terms generated from the user’s web
usage data undergoes the feature population and synonymization. This approach also
considers the user’s social web data to generate terms that match the user’s intent.
A real-world knowledge base is constructed while other approaches use domain
ontology. The Random Forest Algorithm classifies the content, and ACO is applied
to the classified content to select the most relevant tag in response to a particular user
tag. Experimental results reveal an average accuracy and F-measure of 91.79% and
91.74%, respectively, higher than the other baseline approaches.
References
1. De Caro P, Pini MS, Sambo F (2016) A tag-based recommender system. In: Intelligent
autonomous systems 13. Springer, Cham, pp 1049–1061
2. Dorigo M, Stützle T (2019) Ant colony optimization: overview and recent advances. In:
Handbook of metaheuristics. Springer, Cham, pp 311–351
3. Dhandi M, Chakrawarti RK (2016) A comprehensive study of web usage mining. In: 2016
symposium on colossal data analysis and networking (CDAN). IEEE, pp 1–5
4. Anitha V, Isakki P (2016) A survey on predicting user behavior based on web server log files in a
web usage mining. In: 2016 international conference on computing technologies and intelligent
data engineering (ICCTIDE 2016). IEEE, pp 1–4
5. Wei J, He J, Chen K, Zhou Y, Tang Z (2017) Collaborative filtering and deep learning based
recommendation system for cold start items. Expert Syst Appl 69:29–39
6. Hameed MA, Al Jadaan O, Ramachandram S (2012) Collaborative filtering based recommen-
dation system: a survey. Int J Comput Sci Eng 4(5):859
7. Subramaniyaswamy V, Logesh R, Chandrashekhar M, Challa A, Vijayakumar V (2017) A
personalised movie recommendation system based on collaborative filtering. Int J High Perform
Comput Networking 10(1–2):54–63
8. Jiang L, Cheng Y, Yang L, Li J, Yan H, Wang X (2019) A trust-based collaborative filtering
algorithm for E-commerce recommendation system. J Amb Intell Hum Comput 10(8):3023–
3034
9. Fu M, Qu H, Yi Z, Lu L, Liu Y (2018) A novel deep learning-based collaborative filtering
model for recommendation system. IEEE Trans Cybern 49(3):1084–1096
10. Jain KN, Kumar V, Kumar P, Choudhury T (2018) Movie recommendation system: hybrid
information filtering system. In: Intelligent computing and information and communication.
Springer, Singapore, pp 677–686
11. Kawasaki M, Hasuike T (2017) A recommendation system by collaborative filtering including
information and characteristics on users and items. In: 2017 IEEE symposium series on
computational intelligence (SSCI). IEEE, pp 1–8
12. Davagdorj K, Park KH, Ryu KH (2020) A collaborative filtering recommendation system
for rating prediction. In: Advances in intelligent information hiding and multimedia signal
processing, pp. 265–271. Springer, Singapore
13. Deepak G, Santhanavijayan A (2020) OntoBestFit: a best-fit occurrence estimation strategy
for RDF driven faceted semantic search. Comput Commun 160:284–298
564 D. Surya et al.
14. Pushpa CN, Deepak G, Kumar A, Thriveni J, Venugopal KR (2020) OntoDisco: improving web
service discovery by hybridization of ontology focused concept clustering and interface seman-
tics. In: 2020 IEEE international conference on electronics, computing and communication
technologies (CONECCT). IEEE, pp 1–5
15. Kumar A, Deepak G, Santhanavijayan A (2020) HeTOnto: a novel approach for concep-
tualization, modeling, visualization, and formalization of domain centric ontologies for heat
transfer. In: 2020 IEEE international conference on electronics, computing and communication
technologies (CONECCT). IEEE, pp 1–6
16. Deepak G, Kasaraneni D (2019) OntoCommerce: an ontology focused semantic framework
for personalised product recommendation for user targeted e-commerce. Int J Comput Aided
Eng Technol 11(4–5):449–466
17. Gulzar Z, Leema AA, Deepak G (2018) PCRS: personalized course recommender system based
on hybrid approach. Procedia Comput Sci 125:518–524
18. Deepak G, Teja V, Santhanavijayan A (2020) A novel firefly driven scheme for resume parsing
and matching based on entity linking paradigm. J Discret Math Sci Cryptogr 23(1):157–165
19. Haribabu S, Sai Kumar PS, Padhy S, Deepak G, Santhanavijayan A, Kumar N (2019) A
novel approach for ontology focused inter-domain personalized search based on semantic set
expansion. In: 2019 fifteenth international conference on information processing (ICINPRO),
Bengaluru, India, pp 1–5. https://doi.org/10.1109/ICInPro47689.2019.9092155
20. Deepak G, Kumar N, VSN Sai Yashaswea Bharadwaj G, Santhanavijayan A (2019) OntoQuest:
an ontological strategy for automatic question generation for e-assessment using static and
dynamic knowledge. In: 2019 fifteenth international conference on information processing
(ICINPRO). IEEE, pp 1–6
21. Santhanavijayan A, Naresh Kumar D, Deepak G (2021) A semantic-aware strategy for auto-
matic speech recognition incorporating deep learning models. In: Intelligent system design.
Springer, Singapore, pp 247–254
22. Deepak G et al (2019) Design and evaluation of conceptual ontologies for electrochemistry as
a domain. In: 2019 IEEE international WIE conference on electrical and computer engineering
(WIECON-ECE). IEEE
23. Deepak G, Priyadarshini JS (2018) Personalized and enhanced hybridized semantic algorithm
for web image retrieval incorporating ontology classification, strategic query expansion, and
content-based analysis. Comput Electr Eng 72:14–25
24. Sumathi G, Akilandeswari J (2020) Improved fuzzy weighted-iterative association rule based
ontology postprocessing in data mining for query recommendation applications. Comput Intell
36(2):773–782
25. Bt HK, Vibha L, Venugopal KR Personalized recommendation of web pages using group
average agglomerative hierarchical clustering (GAAHC)
26. Gopalakrishnan T, Sengottuvelan P, Bharathi A, Lokeshkumar R (2018) An approach to
webpage prediction method using variable order Markov model in recommendation systems.
J Internet Technol 19(2):415–424
MKSTD: Multiple Knowledge Sourcing
Approach for Topic Directories
Construction and Linking Web Pages
Using Fusion Semantics
1 Introduction
The existing structure of the Web is tending towards “Web of Knowledge” and Web
2.0 is converging into the Semantic Web 3.0. The Web contains a lot of hierarchical
data that is organized into a classified structure, but it is highly disintegrated in
several pockets and not structured together. This creates a distinction between the
integrative nature of the Web information. There is always a necessity to integrate
several relevant contents in the World Wide Web based on topical hierarchy and
link the entities in the Web together to facilitate Semantic Search. The structure of
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 565
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_51
566 G. Deepak and A. Santhanavijayan
the Web is beyond one’s imagination as the entire structure of the Web cannot be
visualized together owing to the millions of zettabytes of data. The size of the indexed
Web alone is approximated to 5.59 billion pages. The only strategy to align the web
pages with the underlying semantic layer of the web is to generate topic directories
and facilitate a link among the topics in the Web to furnish Open Linked Data for
individual topics and anchor the indexed web pages to these topics based on the nature
and the structure of the World Wide Web. Dynamic Generation of Topical Knowledge
from several highly significant data repositories like the Knowledge Bases, Crowd-
Sourced Intelligence, and the traditional Web Directories for facilitating linking of
the web pages is the key towards knowledge-centric semantic search. Most of the
existing approaches either use static Ontological Models as knowledge or use one of
the standard Knowledge bases or a combination of the two. A few other approaches
did not make use of Ontological Models but depended on Web Directories, and a
static knowledge thesaurus. Labels and Embeddings [1] have also played a vital role
in semantic and language processing applications. However, since the present Web
has a facility of Generating Dynamic Knowledge by aggregation of these sources of
Knowledge, a new paradigm for semantic search with increased knowledge density
and a higher degree of relevance is the key towards a Knowledge Fusion Paradigm
in facilitating Semantic Search.
Motivation: Due to the density of the Semantic Web, there is a presence of a large
number of topics that are related. There is always a need for generating topic hier-
archies and leveraging the existing web pages to these topic hierarchies. This can
facilitate Semantic Search by accelerating the quality of recommendations from the
linked data standards of the Semantic Web 3.0. The existing structure of the Web
is also organized to a certain extent, but useful data resources are either separately
stored in Knowledge Bases or are distributed as a collective collection in domain-
based websites or blogs or the traditional and the forgotten Web Directories. An
approach to integrate all these contents to generate Topic Hierarchies and anchor
web pages to them is the need of the hour.
2 Related Work
Sun et al. [1] have proposed a scheme for leveraging label and word embedding
for the semantically driven discovery of web services in a sparse environment. The
approach used label based neutral topic model for web services retrieval based on
categories of web services. Qi et al. [2] have put forth a keyword-driven strategy for
test model generation for web applications by discovering relevant states and transi-
tions. Delpeuch et al. [3] have proposed a simple scheme for Named Entity linking for
aggregating contents from Wikidata to facilitate lightweight Entity linking. Carlos
et al. [4] have introduced an LDA with a probabilistic topic model for Topics
Discovery in Semantic Markup Literature. Sahiti et al. [5] have proposed a strategy
to link lecture slides to facilitate navigation such that a Web of Slides using the simi-
larity of slides as a core concept. Patricia et al. [6] proposed a four-stage approach for
linguistic Linked Open Data generation for the Legal Domain based on identification,
generation of language resources and further converting into RDF, and finally linking
the resulting terminological resources. El Vaigh et al. [7] have put forth a strategy of
linking Entities based on the semantics of the knowledge base. The interrelationship
semantics based on weighting the relations based on the RDF KB structure has been
proposed. Jang et al. [8] have proposed an approach for the automatic construction
of Taxonomy from the Text-Rich Network based on an instance level selection and
term embedding based on the semantics and node granularity. The approach also
integrates clustering to yield subtopics for the taxonomy node. Venugopal et al. [9]
have proposed a Levenshtein Similarity Weight computation scheme for mapping
web pages with topic directories by also incorporating a hashing scheme for linking
web pages with topic hierarchies. Leena et al. [10] have proposed the MA algorithm
for the generation of related clustered hierarchy for the Semantic Web for Keyword
Search which takes the level information and the entity value for fetching the clusters
which match the input keyword where a concise summary tree is constructed. Deepak
et al. [11] have proposed the OntoBestFit which predominantly uses the Simpson’s
Diversity Index with the EnAPMI measure in an environment of static and dynamic
Ontologies for yielding web pages based on detailed ontology semantics. Giri et al.
[12] have proposed the OntoYield which is a semantic similarity-driven contextual
method for the generation of ontologies where the structure of the relationships is
preserved. In [13–16] several Ontology-driven approaches in several applications
have been proposed.
568 G. Deepak and A. Santhanavijayan
The architecture of the web pages to topic mapper has been proposed which is a
dynamic web page to a topic mapper that mainly focuses on linking the real-world
knowledge with the existing web pages based on the indices of the web pages which
ensures the formulation of a reasonable Semantic Web. Since the World Wide Web
is transfiguring into a Web of Data, there is a need for organizing the contents on
the World Wide Web into topic hierarchies such that newer links can be formu-
lated among the existing knowledge bases. The easiest way to do this is to build a
dynamic agent for aggregating terms from various sources and intelligently collect
the crowd intelligence based on the topic-based directory for browsing the data on
the World Wide Web. The users’ intelligence must be integrated into the existing
topic directories to create newer links or even loops on the contents of the World
Wide Web.
The architecture of the proposed Semantic Mapper is based on the Diversification
of the Terms by integrating real-world knowledge into the proposed approach. In
the proposed approach the Linked Open Data is generated, and the web pages are
mapped based on their indexing and contents. The hierarchical organization of the
topics is checked pertaining to seven-stage levels of relevance when looked on with
the parent term. Also, the anchorage of web pages with the terms that have been
discovered is evaluated for its relevance. The architecture of the proposed MKSTD is
depicted in Fig. 1 where an array of knowledge sources is integrated for a hierarchical
topic generation. The integration of static local Domain Ontologies with real world
knowledge bases namely the DBPedia, YAGO, NELL, Wikidata, DebatePedia, and
even quite others has been common. However, in the Semantic Web Mapper, apart
from these knowledge bases and the Domain Ontologies modeled, the sessional data
for every user query from several profiles of web usage data has been considered.
Apart from the individual user-click information of the selective users, certain old-
fashioned web directories are crawled and are made use for experimentation, apart
from the index words or labels in the Web dataset.
The Web Directories are a useful source of information and serve as a massive
knowledge base. The experimentation queries were used to crawl the contents from
these old-fashioned directories which serve as a ready knowledge base that has been
forgotten to be an intrinsic part of the web search engines. However, for populating
the individuals to the entities for the facts in the knowledge bases, the Web Directo-
ries are quite essential. Moreover, when the entities from Web Directories have been
anchored to the ontologies and knowledge bases, the density of information over a
spear of relevance increases, thereby the probability of knowledge for a specific user
query or a topic increases enhancing the diversity of results that are recommended
for a specific topic or a user query. The Web Directories that have been encompassed
are the Business Bureau, BOTW, Yelp, Foursquare, and HotFrog. The interface of
the user-intent and query words is integrated through an Agent into these Web Direc-
tories and their web repositories are crawled using a localized crawler. Owing to
MKSTD: Multiple Knowledge Sourcing Approach for Topic Directories … 569
the incorporation of these directories for the user intents and queries, the amount of
information density of facts over a specific topic is quite high.
A topic-based agent is employed for crawling domain adherent knowledge from
Knowledge Bases and stored as a Subject, Predicate, and Object from the series of
Web Directories. However, SPARQL driven agents have been employed to query the
Knowledge Bases based on the user intents and user query topics. Apart from this,
the experimentation topics were given to several users’ and their browsing history
was collected, and a topic mapping based on the relevance of terms among the
user intents have been mapped as an open topic directory and are mainly used for
leveraging the web pages onto these topics. The relevant items are selected based
on the semantic similarity computation which is achieved using the Jaccard Index
[17, 18] and the Simpson’s Diversity Index [19] to yield the diversified entities. The
570 G. Deepak and A. Santhanavijayan
reason for using the Jaccard Index is mainly owing to its ease of use and its efficient
operability between the sample sets. The Simpson’s Diversity Index works well
especially in an environment where there is a high degree of diversity in entities that
are relevant. In this case, owing to the heterogeneity of Knowledge Resources such
as the Web Directories, Static Domain Ontologies, Real-World Knowledge Bases,
Crowdsourced Topic Hierarchies in both the concept selection and the selection of
individuals.
4 Implementation
The proposed system is implemented in JAVA and the SPARQL Endpoints are
embedded in a JADE driven multiagent. The Domain Ontologies are modified using
Web Protégé and OntoCollab. The knowledge is modeled using the RDF which
is expressed in N-Triple format for storing the crowdsourced knowledge as Facts.
The Web directories are also accessed using the Multiagent JADE system. Stanford
CoreNLP has been encompassed for pre-processing and for all language processing
tasks. The proposed MKSTD algorithm has been depicted in Table 1 for Linked
Open Data Generation and linking of web pages with the generated topic hierarchy.
The experimentations were carried out for the Bookshelf from the Gutenberg Project
categorizations (https://www.gutenberg.org/ebooks/bookshelf/). The categorizations
in the Gutenberg Project Bookshelf were used as a topic directory for generating topic
hierarchies and thereby anchoring web pages to the topic directories. Standard metrics
such as the Precision, Recall, Accuracy, F-Measure, nDGC (Normalized Discounted
Cumulative Gain), and the False Discovery Rate (FDR) was used for evaluating the
performance of the proposed approach. The nDGC computes the diversity in the
furnished results. However, the other metrics compute the relevance of the results
yielded by the proposed approach.
It is evident from Table 2 that the proposed MKSTD yields an average Precision
of 94.89%, an average Recall of 97.04%, an average Accuracy of 95.97% with
a very low FDR of 0.06, and an nDGC score of 0.91. The proposed MKSTD is
baselined with two approaches namely the MPC-LSW [9] and the MA Algorithm
[10] and several variations by excluding the integrated approaches are used as a
standard for comparing the efficacy of the approaches integrated into the MKSTD.
The proposed MKSTD has a 6.75% of higher precision, 6.9% of higher recall, than
the MPC-LSW, and 6.83% of accuracy higher than that of MPC-LSW. MKSTD has
achieved an increase in the Precision, Recall, and Accuracy of 11.71%, 11.16%, and
11.84% respectively when compared to the precision, recall, and accuracy of the MA
Algorithm. When the Web Directories were removed from the proposed MKSTD,
MKSTD: Multiple Knowledge Sourcing Approach for Topic Directories … 571
Table 1 Proposed MKSTD algorithm for linked open data generation and mapping of web pages
Input: The query Q (Query Words can be changed to terms) which is input as a
keyword or a set of multi-keyword, Query Word Based Classified Knowledge
Resources namely the Directories, Ontologies, Real-World Knowledge Bases, and
Crowdsourced Topic Hierarchies.
Output:
Begin
Step 1: The Query Q if multiword is preprocessed by lemmatization, stemming, and
Stop word removal.
Step 2: for each qw in Q
Set Onto← Derive EntitiesKnowledge from the Static Domain Ontologies
end for
Step 3: while (Onto.next()!=NULL)
Jd←JaccardSimilarity(Onto(i), SPARQLEndpoints(Kb)
Sd← SimpsonsDiversityIndex(Onto(i),Kb)
i++;
return Fact←Jd U Sd
end while
Step 4: while (Fact.next()!=NULL)
Jd←JaccardSimilarity(Fact(i), CrowdEmbeddings)
Sd← SimpsonsDiversityIndex(Fact(i), CrowdEmbeddings)
i++;
return EnrichedFact←Jd ∩ Sd
end while
Step 5: Using an agent, query the entities in the EnrichedFact to Web Directories and
populate the individuals from the list of Web Directories to Generate OLData.
Step 6: Based on the Computation of the Jaccard Similarity and the Simpsons
Diversity Index, link the web page URLs with the OLData entities, based on
the index words and the content words of the page.
Step 7: Recommend the OpenLinkedData OLData, the Mappings of the Topic
Hierarchies to the user.
end
the accuracy receded by 7.88%. Similarly, when the Diversity Index was removed
from MKSTD, the accuracy decreased from 95.97 to 90.37%. When there was only
a single knowledge base in the proposed approach, the overall accuracy decreased by
6.76%. Since the Accuracy is derived from Precision, and Recall, the direct decrease
in the accuracy signifies a considerable decrease in the percentages of Precision, and
Recall respectively, and a corresponding increase in the FDR value.
A powerful approach of a lethal combination of a standard Semantic Similarity
measure and a Diversity Index, for generating the Linked Open Data as topic hier-
archies, and leveraging the web pages to the topic that happens based on the index
words or the content words in the web pages. Apart from this, there is a large number
of real-world entities injected into the approach from several heterogeneous sources
namely the DBPedia, YAGO, NELL, Wikidata, and the DebatePedia. The Knowl-
edge Base resources form a major part of the TBOX of the Ontologies. The array of
Web Directories populates the individuals to the conceptual schema, thereby trans-
figuring this approach into a perfect knowledge-driven approach for generating topic
hierarchies and leveraging web pages based on the topic. Moreover, the usage of the
Diversity Index discovers the terminologies that are relevant and novel. The usage of
Jaccard Similarity ensures that there is no divergence with respect to the relevance
of the web pages. The F-Measure percentage Vs the Number of Recommendations
is depicted in Fig. 2.
The proposed MKSTD has the lowest FDR of 0.06 and the highest nDCG score of
0.91 when compared with the baseline approaches and other variants when certain
important methodologies were eliminated. The MKSTD has an FDR of 0.06 less
than MPC-LSW and 0.1 nDCG more than the MPC-LSW. The MKSTD has FDR of
0.11 less than that of the MA Algorithm and has an nDCG of 0.13 higher than the
MA Algorithm. Since all the other variants of the strategies in MKSTD has one of
the important techniques eliminated from its hybridization, all the other variations
have lower values of Precision, Recall, Accuracy and nDCG when baselined with
the MKSTD. The FDR of MKSTD is much lower than the other variants. The reason
for the high values of the nDCG in the MKSTD is mainly due to the incorporation of
a large range of heterogenous knowledge sources ranging from standard knowledge
bases that hold facts to the Web directories, and the infusion of the Simpson’s Diver-
sity Index. The FDR is quite low mainly due to the regulation of using a combined
semantic similarity measure, the Jaccard Similarity with the Simpson’s Diversity
Index. Owing to the same reason, the F-Measure Vs Number of Recommendations
for the MKSTD is higher than the baseline approaches and other variants of the
MKSTD when one of the important hybridizations has been eliminated.
6 Conclusions
The MKSTD framework for mapping and integrating knowledge from several hetero-
geneous sources for topic hierarchy generation and linking of web pages has been
proposed. The approach integrates knowledge from a series of knowledge bases,
crowd intelligence, and a number of traditional Web directories. The Linked Open
Data is generated by linking the topic hierarchies using a combination of the Jaccard
Similarity and Simpson’s Diversity Index. Also, the anchorage of the web pages with
the topic hierarchies in the Linked Open Data is achieved by the combination of the
Jaccard Similarity with the Simpson’s Diversity Index. The MKSTD is one of the
first systems to have integrated the standard Knowledge Bases with the traditional
web directories to harvest knowledge and incorporate the same in an environment of
Semantic Search. The MKSTD has achieved an average accuracy of 95.97% with an
nDCG of 0.91 and a very low FDR of 0.06 for the Gutenberg Project categorizations
which proves the efficacy of relevant knowledge integration into an environment of
semantic search.
References
1. Sun C, Lv L, Tian G, Wang Q, Zhang X, Guo L (2020) Leverage label and word embedding
for semantic sparse web service discovery. Math Probl Eng
2. Qi XF, Hua YL, Wang P, Wang ZY (2019) Leveraging keyword-guided exploration to build
test models for web applications. Inf Softw Technol 111:110–119
574 G. Deepak and A. Santhanavijayan
3. Delpeuch A (2019) Opentapioca: lightweight entity linking for wikidata. arXiv preprint arXiv:
1904.09131
4. Montenegro C, Navarrete R (January 2020) Improving core topics discovery in semantic
markup literature: a combined approach. In: Proceedings of the 3rd International Conference
on Software Engineering and Information Management, pp 237–243
5. Labhishetty S, Pei K, Boughoula A, Zhai C (2019) Web of slides: automatic linking of
lecture slides to facilitate navigation. In: Proceedings of the Sixth (2019) ACM Conference on
Learning@ Scale, pp 1–4
6. Martín-Chozas P, Montiel-Ponsoda E, Rodríguez-Doncel V (2019) Language resources as
linked data for the legal domain. Knowl Law Big Data Age 317:170
7. El Vaigh CB, Goasdoué F, Gravier G, Sébillot P (September 2019) Using knowledge base
semantics in context-aware entity linking. In: Proceedings of the ACM Symposium on
Document Engineering 2019, pp 1–10
8. Shang J, Zhang X, Liu L, Li S, Han J (April 2020) Nettaxo: automated topic taxonomy
construction from text-rich network. In: Proceedings of the Web Conference 2020, pp
1908–1919
9. Venugopal KR, Srikantaiah KC (2020) Construction of topic directories using Levenshtein
similarity weight. In: Web Recommendations Systems. Springer, Singapore, pp 67–86
10. Giri LG, Mathur A, Manjula SH, Venugopal KR, Patnaik LM (2014) MA Algorithm to generate
semantic web related clustered hierarchy for keyword search. Int J Recent Trends Eng Technol
11(1):222
11. Deepak G, Santhanavijayan A (2020) OntoBestFit: a best-fit occurrence estimation strategy
for RDF driven faceted semantic search. Comput Commun 160:284–298
12. Giri GL, Deepak G, Manjula SH, Venugopal KR (2018) OntoYield: a semantic approach
for context-based ontology recommendation based on structure preservation. In: Proceedings
of International Conference on Computational Intelligence and Data Engineering. Springer,
Singapore, pp 265–275
13. Kumar A, Deepak G, Santhanavijayan A (July 2020) HeTOnto: a novel approach for
conceptualization, modeling, visualization, and formalization of domain centric ontologies
for heat transfer. In: 2020 IEEE International Conference on Electronics, Computing and
Communication Technologies (CONECCT). IEEE, pp 1–6
14. Deepak G, Priyadarshini S (2016) A hybrid framework for social tag recommendation using
context driven social information. Int J Soc Comput Cyber-Phys Syst 1(4):312–325
15. Deepak G, Teja V, Santhanavijayan A (2020) A novel firefly driven scheme for resume parsing
and matching based on entity linking paradigm. J Discrete Math Sci Cryptogr 23(1):157–165
16. Haribabu S, Sai Kumar PS, Padhy S, Deepak G, Santhanavijayan A, Kumar DN (2019) A
novel approach for ontology focused inter-domain personalized search based on semantic set
expansion. In: 2019 Fifteenth International Conference on Information Processing (ICINPRO),
Bengaluru, India, pp 1–5. https://doi.org/10.1109/icinpro47689.2019.9092155
17. Jaccard P (1901) Étude comparative de la distribution florale dans une portion des Alpes et des
Jura. Bulletin de la Société vaudoise des sciences naturelles 37:547–579
18. Jaccard P (1912) The distribution of the flora in the alpine zone. New Phytol. 11(2):37–50.
https://doi.org/10.1111/j.1469-8137.1912.tb05611.x
19. Simpson EH (1949) Measurement of diversity. Nature 163(4148):688
A Fuzzy Near Neighbors Approach
for Arabic Text Categorization Based
on Web Mining Technique
Mounir Gouiouez
Abstract Nowadays, the quantity of textual content available online has experienced
such a colossal increase. Hence, the need for a system to investigate this content data
is mandatory. In this concern, Text Categorization (TC) highlights many performance
methods and techniques to analyze, explore and classify various types of documents.
This study consists of two main steps. First, we extract terms from text documents
using Fuzzy Near Neighbors (FNN) with web-based mining techniques algorithm.
Second, we identify documents according to a particular form of similarity based on
combining all Arabic encyclopedic dictionaries using clustering algorithms. In this
article, Fuzzy C-Means (FCM) as a clustering algorithm is used to perform the preci-
sion of documents’ classification. This work suggests Arabic TC based on a multi-
lingual encyclopedic dictionary (Arabic WordNet, OMW, Wikipedia, OmegaWiki,
Wictionary, and Wikidata). To evaluate the efficacy of TC approach with FNN and
FCM, an experimental study using a real-world dataset is carried out. The results
of the present study indicate that proposed approach outperforms the traditional one
and produces good results.
1 Introduction
M. Gouiouez (B)
Psychological, Social and Cultural Studies Laboratory, Sidi Mohamed Ben Abdellah University
(USMBA), Fez, Morocco
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 575
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_52
576 M. Gouiouez
texts written in the Arabic language. Developing TC systems for Arabic documents
is a difficult activity due to the complicated and rich nature of the Arabic language.
Text Classification as described by Baraa et al. [1] in 2014 is a series of supervised
learning tasks undertaken to structure a collection of documents using a variety of
different methods, including distance-based, tree-based decision-making, Bayesian
classification, SVM … etc. TC techniques are used in many disciplines, including
mail filtering, irrelevant filtering messages, document archives, and automatic data
indexing of scientific articles. In theory, the standard text categorization process
consists of four stages, each of which includes many steps.
These stages provide as first the pre-processing phase of the document. This stage
involves several actions to prepare the content of documents that are seen by the
classification model. It tokenizes the text, removes stop-word, and weights terms.
The second stage of TC is the extraction of terms, follow by the last step of mapping
and disambiguation.
This study proposes a generic approach for extracting words using Fuzzy
near neighbors with web-based mining techniques. This approach identify docu-
ments according to a particular form of similarity based on combining all Arabic
encyclopedic dictionaries using clustering algorithms.
Fuzzy C-Means, as a clustering algorithm, is used to perform the approach of this
research. This study adapts an experimental design using real-world dataset.
The objective of this article is twofold. First, it seeks to investigate the ability
of constructing a fuzzy logic theoretical model for Arabic text categorization that
provides appropriate results in relation with other different classifiers. Second, it
aims to suggest an enhanced version of the easy reduce way to create various high
reduction based on all Arabic encyclopedic dictionaries.
This paper is organized in this manner. Section 2 presents a thorough definition
of the key concepts used in this article. Section 3 gives an overview of Arabic Text
Categorization. The design of the approach used in this study is provided in Sect. 4
with its different steps. Section 5 presents the experimental results. Finally, Sect. 6
provides further directions to the TC problem using the Fuzzy logic.
2 Preliminaries
According to [2], the web data mining process consists of three steps: pre-processing
data, discovering patterns, and analyzing patterns. The first step, pre-processing, is
concerned with converting web data into an easily evaluated data. The second stage
of web data mining deals with the extraction of special knowledge and patterns
from labeled data using mining and machine learning algorithms [3]. The last step,
patterns’ analysis, copes with collecting similar patterns from the appropriate ones.
Web data Mining is the process of data mining methods for useful information
from the web and its trends of use. Web mining helps to increase the control of the web
search by categorizing web documents and recognizing web content. It is used for
search engine, to predict behavior of users, and for a specific website encyclopedic
dictionary. Web Data Mining and knowledge exploration are continually growing
for researchers [4]. Web data mining generally contains two categories: web services
Mining and Web Structure Mining. Web services mining is the request of extracting
beneficial data from the content of web documents, while web structure mining is
an application for the discovery of organizational features from web content [5].
FCM [6] is an unsupervised learning algorithm introduced by Dunn in 1973 [7] and
modified by Bezdek in 1987 [8]. It is a clustering method that split one corpus of data
to their correspond clusters (two or more). This methodology is utilized for selecting
adequate information. This FCM is founded on optimizing the following function:
2
N
C
Dinj |xi − cj| (1)
i=1 j=1 1≤m≤∞
N
Dinj xi
1 i=1
Di j = c 2 cj = (2)
|xi −c j | n−1 N
xi −ck Dinj
k=1 i=1
578 M. Gouiouez
If maxi j |Di(k+1)
j − Di(k)
j | ≤ ε, where ε is a condition criterion between 0 and 1,
the iteration loop stopped. K number of iteration steps. The algorithm is composed
of the following steps:
Step 1: Initialize D = Di j matri x, D(0)
Step 2: estimate the center vectors C(k) = c j with D(k) by:
N
Dimj xi
i=1
cj =
N
Dimj
i=1
1
Di j = c 2
|xi −c j | m−1
xi −ck
k=1
During the last years, numerous studies have been conducted on the importance of
Arabic TC. These studies need are briefly discussed in order to provide empirical
evidence for the ongoing research. For instance, in 2006, many research studies used
the K-nearest neighbor technique to categorize data in Arabic TC [9]. Duwairi et al.,
in [10] listed the main classifiers in their article that have shown great effectiveness in
Arabic text processing- Naïve Bayes (NB), K-nearest neighbor (K-NN) and distance-
based algorithms. Then, Mesleh [11] suggests a method that use support vector
machine (SVM).
In 2011, several studies on Arabic text categorization were conducted using
classical methods used before but for different data resource and with different
pre-processing methods [12].
In 2012, the authors of [13] introduced a neural network-based approach with the
learning vector quantization (LVQ) algorithm and the Kohonen self-organizing map
(SOM) to discovery similarities in text for categorization. In 2015, a comparative
study of several approaches of Arabic TC presented in [14]. Similarly, Al-Radaideh
and Al-Khateeb used a hybrid classifier to classify Arabic documents. Table 1 shows
the most popular contributions in Arabic Text Categorization.
However, almost all works have taken into account the specific resource based
on a particular model and studies the relationship between term to documents or
documents to categories. In fact, in our model we present a design that calculates
A Fuzzy Near Neighbors Approach for Arabic Text Categorization ... 579
the semantic similarity between two terms based on multiple resources before the
extracting phase and organizing the datasets in cluster prior to classification.
4 Our Approach
4.1 Methodology
The current study is based on the idea that semantically similar terms are used in
similar contexts in documents. We also assume that the set of words that co-occur in
580 M. Gouiouez
Fig. 1 Illustration of our design with fuzzy near neighbors and clustering algorithm
a snippet with a target term to constitute a common meaning for the occurrence of
the target word with K Fuzzy near neighbors.
The approach suggested in this research is composed of two stages as indicated in
(Fig. 1). The first stage relates to the learning phase. It consists of generating a new
text representation based on mapping terms with their associated concept using K
Fuzzy near neighbors. The second stage relates to the classification phase. It consists
in weighting categories, and calculating the profile of the document to be classified
using Clustering algorithm for the creation of a multi classifier system.
With this approach, we aim for two things: the increase of similarity of extract
terms via the web mining and the reduction of the size of documents through the
Clustering algorithm.
A Fuzzy Near Neighbors Approach for Arabic Text Categorization ... 581
K Fuzzy Near Neighbors. This type of classifier takes advantage of fuzzy logic in
distance calculation used for the selection of the K nearest neighboring. In [15], the
authors define this fuzzy measure as follows:
Fe
−1.0
d(x, R)
u(x) = 1 + (3)
Fd
With: d(x, R) the distance between the unknown term x and the sample R of resource.
Fd and Fe are two positive constants used to control Fuzzy term in the previous
calculation.
DicƟonnary
0
Different companies’ data has been selected to conduct the present study. The data
selected include small and large corpus with few and many categories. The companies
from which the data is collected include the CCA datasets, the BBC-Arab datasets,
and the EASC datasets (Fig. 3).
The assessment of the performance of the classification algorithm for classifying data
into the appropriate category shall be carried out using a number of mathematical
rules, such as recall, precision and F-measure, which are described as follows:
In this section, we introduce and evaluate experimental data using the suggested
method (Text Categorization algorithms with SVM and NB techniques). In so doing,
each corpus is divided into two segments (80% for training and 20% for testing). For
testing, a computer with 8 GB RA processor is used.
The assessment of the performance of the classification algorithm for classifying
data into the appropriate category is carried out using a number of mathematical
rules, such as recall, precision and F-measure, which are described as follows.
TP TP
Precision = And Recall = (4)
TP + FP TP + FN
Table 2 Comparison
DataSet EASC CCA BBC-Arabic
between SVM and NB Using
FNN and FC Methods SVM 92 94,5 89,4
NB 84,2 83,25 87,2
SVM + Fuzzy logic clustering 92,07 95,52 89,89
NB + Fuzzy logic clustering 83,37 85,01 85,34
In this paper, all the methods discussed organize the features according to their
importance to the category. The top ranking features from each category are then
chosen and represented to the classification algorithm. Consequently, we obtain a
new form of document with a new dimension.
The results showed in Table 2 disclose that Fuzzy logic clustering almost always
performs text classification than any of the other learning algorithms in different
corpora. In regard to the results of all classifiers, Table 2 has revealed that SVM
combined by Fuzzy logic clustering method for feature reduction are typically better
than NB.
6 Conclusion
This study has approached the problem of text representation for Arabic text docu-
ments. It has introduced a novel approach using Fuzzy Near Neighbors (FNN) with
web-based mining techniques algorithm. Similarly, this paper has identified docu-
ments according to a particular from of similarity based on the combination of all
Arabic encyclopedic dictionaries using clustering algorithms. The results of this
study have shown that the suggested method increase the classification of Arabic
documents better than the classical techniques.
References
1. Baraa S, Nazlia O, Zeyad S (2014) An automated Arabic text categorization based on the
frequency ratio accumulation. Int Arab J Inform Technol 11(2):213–221
2. Kumar GD, Gosul M (2011) Web mining research and future directions. In: Wyld DC,
Wozniak M, Chaki N, Meghanathan N, Nagamalai D (eds.) Advances in Network Security
and Applications. CNSA
3. Tharwat A (2019) Parameter investigation of support vector machine classifier with kernel
functions. Knowl Inf Syst 61:1269–1302
4. Ali N, Neagu D, Trundle P (2019) Evaluation of k-nearest neighbour classifier performance
for heterogeneous data sets. SN Appl Sci 1:1559
5. Chen J, Dai Z, Duan J et al (2019) Improved Naive Bayes with optimal correlation factor for
text classification. SN Appl Sci 1:1129
584 M. Gouiouez
6. Gavade AB, Rajpurohit VS (2020) Sparse-FCM and deep learning for effective classification
of land area in multi-spectral satellite images. Evol Intel
7. Schroder HW, Welling H, Wellegehausen B (1973) Appl Phys 1:343–348
8. Bezdek JC (1987) Analysis of Fuzzy Information, vol 1, 3. CRC Press, Boca Raton
9. Al-Radaideh QA, Al-Abrat MA (2019) An Arabic text categorization approach using term
weighting and multiple reducts. Soft Comput 23:5849–5863
10. Chantar H, Mafarja M, Alsawalqah H, et al (2019) Feature selection using binary grey wolf
optimizer with elite-based crossover for Arabic text classification. Neural Comput Appl
11. Mesleh A, Ghassan G (2008) Support vector machine text classification system: using ant
colony optimization based feature subset selection. Comput Eng Syst
12. Razavi AR, Gill H, Åhlfeldt H, Shahsavar N (2005) A Data Pre-Processing Method To Increase
Efficiency And Accuracy In Data Mining. In: Miksch S, Hunter J, Keravnou ET (eds.) Artificial
Intelligence in Medicine. AIME 2005. Lecture Notes in Computer Science, vol 3581. Springer,
Heidelberg
13. Biehl M (2012) Admire LVQ—adaptive distance measures in relevance learning vector
quantization. Künstl Intell 26:391–395
14. Al-Radaideh Q, Al-Khateeb S (2015) An associative rule-based classifier for Arabic medical
text. Int J Knowl Eng Data Min 3(3–4):255–273
15. Kramer O (2013) K-nearest neighbors. In: Dimensionality Reduction with Unsupervised
Nearest Neighbors. In: Intelligent Systems Reference Library, vol 51. Springer, Heidelberg
Recommendation System Based
on Semantic Web Approach
Abstract Recently, most countries suffer a decrease the crop productivity due to
weather conditions, so many kinds of research are suggested for growth and enhance
crop productivity with different methodologies and techniques. The Semantic Web
(SW) approach is proposed in this work to build a recommendation system for helping
the farmers to take an adequate decision of recommended fertilizer in fast time
by specific parameters (e.g., temperature, humidity, moisture, soil type, crop type,
nitrogen, potassium, and phosphorus). This proposed approach is based on some
essential technologies such as Ontology, Web Ontology Language (OWL), Resource
Description Framework (RDF), and RDF Query Language (SPARQL). In this work,
our goal is to present another technique to improve the quality of service for farmers
with more intelligence in order to increase their crop productivity. The results showed
that the capability of the recommendation system based on SW was effective.
1 Introduction
The Semantic Web (Web 3.0) is the World Wide Web extension, standardized by
the World Wide Web Consortium (WC3) (Berners-Lee et al. 2001) [1]. This term
represents a field of research as other researches domains (e.g., artificial intelligence,
machine learning, etc.). The purpose of SW is to improve the current World Wide
Web by machine-understandable semantic documents and data on the web with the
fast response of user’s questions thanks to its technologies used. So, the Semantic web
has come to solve the most problems facing the current Web, due to the increasing
the data and information day by day that why it will be more difficult to facilitate its
management in the WWW and can’t interpret its semantic. There are some essential
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 585
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_53
586 S. Belattar et al.
implement the recommendation systems in the agriculture field. In the third section,
we describe our methodologies proposed for the Semantic Web and its technologies.
In the fourth section, the results achieved by applying our approach. Finally, we give
a conclusion and perspectives of our next work.
2 Related Work
During the literature review, many works developed the recommendation systems
in agriculture with different methods and techniques such as; a system of recom-
mendation was implemented to recommend the appropriate fertilizer related to
specific soil fertility, climate, and crop by using the Nutrient Expert (NE) model
is a computer-based decision support system, they calibrated Nutrient Expert for
maize agro-ecological areas in Nigeria, Ethiopia, and Tanzania by generating the
data from 735 trials on-farm between 2015 and 2018. In this study, their objective
was to reduce the crop productivity risk and efficiency of using fertilizer to help the
farmers to increase their crop productivity. The results showed that their proposed NE
model was effective from cost-effective for recommending the appropriate fertilizer
to farms by offering another solution for soil testing (Rurinda et al. 2020) [13]. The
agricultural recommendation system was proposed to overcome the pests affecting
a crop by identifying a pest and giving a suitable treatment. For developing this
system, they based on “Pest in Crop and their Treatment” Ontology (PCT-O) in
order to facilitate the usability of their data for making an intelligent system (Lacasta
et al. 2018) [14]. A crop recommendation system was developed to improve the crop
productivity based on ensembling techniques (e.g., Random Forest, Naive Bayes,
and Linear SVM) of machine learning for recommending the right crop related to
soil specific type. Their experimental results showed that the average accuracy of
classification using ensembling techniques was 99.91% (Kulkarni et al. 2018) [15].
The Semantic web (SW) represents an evolution of the World Wide Web. This term of
SW involves specific techniques recommended by the World Wide Web Consortium
(W3C) that improve the current Web and its uses. These techniques provide a meaning
of the data without worrying about its representation (Robu et al. 2006) [18] and it
can understand by both humans and machines. Additionally, the objective of SW is to
create a global database of connected data on the Web (Laufer 2015) [19]. Thus, the
Semantic Web technologies are Ontology, Resource Description Framework (RDF),
Web Ontology Language (OWL), and SPARQL Query language.
Ontology
Firstly, the concept of ontology is originating from philosophy and it represents a
knowledge base (in the symbolic artificial intelligence sense) (Hitzler 2020) [20]. The
main role of ontology is to define the knowledge representation in a specific domain
(Craig 1998) [21], in order to share, exchange, discover the idea or data and improve
the communication between different agents. The ontology contained different
elements (e.g., individuals, classes, attributes, semantic relations, and events) for
improving the Web search. A simple example of crop ontology representation is
shown in (Fig. 1).
In order to get the results, we based on some technologies tools for building our
recommendation system.
In the development of the graphical interface of our recommendation system
based on Semantic Web (Frontend development), HTML5, JavaScript, CSS, and
technique to dynamically create HTML code (JSP) “Java Server Pages” were used
590 S. Belattar et al.
Fig. 3 Recommendation system design flowchart based on the semantic web approach
to implement the web page. Additionally, the Bootstrap framework was used for
designing the graphical interface.
In the development of our recommendation system based on Semantic Web
(backend development), Eclipse IDE for Enterprise Java Developers and Apache Jena
API framework (for creating the Semantic Web Application) were used to implement
all classes and methods to work with fertilizer ontology created and semantic data
contained in the RDF. For creating our fertilizer ontology we based on data that
already exist on the web; these data consist of the specific parameters for recom-
mending the suitable fertilizer. For adding all our classes we used the OWLClasses
tab (see Fig. 4).
For fertilizer type (e.g., DAP and Urea) we add it by clicking on the individuals tab,
and each type of fertilizer related to specific parameters (e.g., temperature, humidity,
moisture, soil type, crop type, nitrogen, potassium, and phosphorus), for adding
these parameters we used the properties tab by clicking on add data type property.
Besides this, each crop can need 2 types of fertilizer according to the parameters we
have as depicted in the figure above (see Fig. 4), the sugarcane crop needs the DAP
and Urea fertilizer for increasing its productivity but with different parameters. By
applying all previous steps for developing our proposed approach, the interface of
our recommendation system based on the SW is ready to use by farmers (see Fig. 5).
In order to get the recommended fertilizer, farmers must insert all parameters of
temperature, humidity, moisture, nitrogen, potassium, phosphorus, crop type, and
soil type, after that the farmers click on the submit SPARQL Query button (see
Fig. 5).
Our ontology is written in RDF/XML, for that we need the SPARQL Query
language for getting the results. When we click on Submit SPARQL Query, the
query used all parameters inserted by the farmer (see Fig. 6). The SPARQL Query
Recommendation System Based on Semantic Web Approach 591
Fig. 5 Recommendation system based on SW - question and suitable fertilizer answer of a farmer
work help the farmers to take suitable decisions of specific fertilizer type in order to
improve their crop productivity.
Recommendation System Based on Semantic Web Approach 593
5 Conclusion
This paper proposes another technique called the Semantic Web approach for devel-
oping the agricultural recommendation system that could improve crop productivity,
this system recommended the appropriate fertilizer for farmers by specific parame-
ters i.e., temperature, humidity, moisture, soil type, crop type, nitrogen, potassium,
and phosphorus. The Semantic Web represents a solution to work with massive data
and different types of data among them non-quantitative data compared to Artificial
Neural Network (ANN) works only with quantitative data as inputs. In this study, our
objective was to help the farmers by improving the Web search, enhancing efficiency,
and giving the answer in fast time. Additionally, this system facilitates the exchanging
the information of both human and machine thanks to Semantic Web Technologies
and it can updates itself dynamically from the ontologies used that why it named the
intelligent web. The results showed that the recommendation system based on the
SW approach was effective. Our perspective is to improve this system in future work
by creating an overall crop ontology that solves several issues in the agriculture field.
594 S. Belattar et al.
References
1. Berners-Lee T, Hendler J, Lassila O (2001) The semantic web. a new form of web content that
is meaningful to computers will unleash a revolution of new possibilities. Sci Am 284(5):34–43
2. Manola F, Miller E, McBride B (February 2004) RDF primer. W3C recommendation, World
Wide Web Consortium, pp 1–107
3. Hommeaux EP, Seaborne A (January 2008) SPARQL query language for RDF. W3C
recommendation, World Wide Web Consortium
4. Gruber TR (1993) A translation approach to portable ontology specifications. Knowl Acquis
5(2):199–220
5. Lima V, Pellison F, Bernardi F, Carvalho I, Rijo R, Alves D (2019) Proposal of an inte-
grated decision support system for Tuberculosis based on semantic web. Procedia Comput Sci
164:552–558
6. Pellison FC, Lima VC, Lopes Rijo RPC, Alves D (2019) Integrating Tuberculosis data in state
of São Paulo over semantic web: a proof of concept. Procedia Comput Sci 164:686–691
7. Yoshiura VT, Yamada DB, Pellison FC, De Lima IB, Damian IPM, Rijo RPCL, De Azevedo
Marques JM, Alves D (2019) Towards a health observatory conceptual model based on the
semantic web. Procedia Comput Sci 138:131–136
8. Kiselev B, Yakutenko V (2020) An overview of massive open online course platforms: person-
alization and semantic web technologies and standards. Procedia Comput Sci 169:373–379
9. Gomathi B, Suguna S (2020) Semantic web application in e-learning using protege based on
information retrieval. J Adv Res Dyn Control Syst 12(7 Special Issue):1847–1855
10. Zangeneh P, McCabe B (2020) Ontology-based knowledge representation for industrial
megaprojects analytics using linked data and the semantic web. Adv Eng Inform 46(June):
11. Marwaha S (20 January 2016) Ontology Based Expert System. Springer
12. Tomic D, Drenjanac D, Wöber W, Hörmann S, Auer W (2015) Enabling semantic web for preci-
sion agriculture: a showcase of the project agriOpenLink. In: CEUR Workshop Proceedings,
vol 1481, pp 26–29
13. Rurinda J, Zingore S, Jibrin JM, Balemi T, Masuki K, Andersson JA, Pampolino MF,
Mohammed I, Mutegi J, Kamara AY, Vanlauwe B, Craufurd PQ (2020) Science-based deci-
sion support for formulating crop fertilizer recommendations in sub-Saharan Africa. Agric Syst
180(July 2019):
14. Lacasta J, Lopez-Pellicer FJ, Espejo-García B, Nogueras-Iso J, Zarazaga-Soria FJ (2018) Agri-
cultural recommendation system for crop protection. Comput Electron Agric 152(June):82–89
15. Kulkarni NH, Srinivasan GN, Sagar BM, Cauvery NK (2018) Improving crop productivity
through a crop recommendation system using ensembling technique. In: 3rd International
Conference on Computational Systems and Information Technology for Sustainable Solutions,
pp 114–119
16. Drury B, Fernandes R, Moura MF, de Andrade Lopes A (2019) A survey of semantic web
technology for agriculture. Inf Process Agric 6(4):487–501
17. Zhang N, Wang M, Wang N (2002) Precision agriculture a worldwide overview. Comput
Electron Agric 36(2–3):113–132
18. Robu I, Robu V, Thirion B (2006) An introduction to the semantic web for health sciences
librarians. J Med Libr Assoc 94(2):198–205
19. Laufer C (2015) Guia_Web_Semantica, p 133
20. Hitzler P (2020) Semantic web: a review of the field. In: Proceedings of ACM, New York, NY,
USA, p 7
21. Craig E (1998) Routledge Encyclopedia of Philosophy. Routledge, New York
22. McGuinness DL, Van Harmelen F (2004) OWL web ontology language overview. W3C
Recomm 10:2004
23. Gandon F, Krummenacher R, Han S-K, Toma I (2011) The resource description framework
and its schema. Handbook of Semantic Web Technologies (Issue January)
24. Cyganiak R, Wood D, Lanthaler M (2014) RDF 1.1 concepts and abstract syntax. W3C Recomm
25
Recommendation System Based on Semantic Web Approach 595
25. Group WSW (2013) SPARQL 1.1 overview. W3C Recomm W3C
26. Che H (2006) A semantic web primer. J Am Soc Inf Sci Technol 27(8)
Sentiment Analysis of Moroccan Tweets
Using Text Mining
Abstract This paper introduces a framework system that analyzes the data extracted
from the social network Twitter. In our study, we deal with Moroccan tweets written
in Standard Arabic (SA) and Moroccan Dialect (MD). This system allows collecting,
processing and visualizing sentiments that could be extracted from Moroccan user’s
communication. As a training set, 36114 tweets were collected but only 13,550
tweets were found to be valid. Firstly, the tweets are classified into 2 categories: SA
and MD. The tweets belonging to each category are then classified into 4 classes:
positive, negative, neutral or mixed. We have performed several experiments to eval-
uate all the possible combinations of the following weight schemes: N-grams and
stemming terms techniques. The experimental results provide the best scenario for
each classifier and indicate that LSTM classifier using word embedding without
removing Stopwords outperforms CNN classifier as well as the CNN-LSTM classi-
fier (with the same features), and it outperforms SVM with term frequency–inverse
document frequency (TF-IDF) weighting scheme with stemming through Bigrams
feature. Moreover, the obtained results outperformed other comparable related work
results.
1 Introduction
With the advance of the web 2.0, many social media platforms emerged, allowing
people to express their feelings and give their opinions regarding a particular
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 597
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_54
598 M. Garouani et al.
subject such as commercial products, services (e.g. travel agencies, internet service
providers) or events (e.g. elections, new laws) [1]. Their interaction on these plat-
forms generates a huge amount of unstructured data. This latter is a valuable resource
that could be used to improve the decision-making process and acquiring knowledge
of people trends.
For this reason, it is useful to analyze the social network platforms (Facebook
or Twitter) content that have gained great popularity around the world [2]. Given
the number of users and their high level of use, these platforms reflect the mood of
the public as well as their trends regarding the current events. Most of the studies
performed in this field have been conducted on foreign languages especially the
English and the Japanese ones [3]. Nevertheless, researches analyzing the sentiment
of written Arabic language are very limited especially on Moroccan dialect.
This can be explained by three factors: Firstly, processing this language is too
complicated [4]. Secondly, there is a lack of web content regarding this language and
thirdly, no extra resources are available for this language [5]. The lack of resources
is considered as the main reason that impedes the development of natural language
processing tools for Arabic dialects in general and the Moroccan dialect (MD) in
particular. There exist no annotated (labeled) corpora for extracting the sentiments
from Moroccan posts. This is the main reason that motivated us to generate an opinion
corpus for the Moroccan dialect.
Indeed, our work aims to develop a system that analyzes the data extracted from the
social network Twitter. The proposed system allows collecting, process and classify
the polarity of Moroccan tweets according to the category they belong to (Standard
Arabic or Moroccan Dialect Arabic), and according to the class they belong to (posi-
tive, negative, neutral, mixed) and ultimately provide statistics for all the regions of
Morocco as well as their dominant sentiments.
In our work we conduct a study to discover the performance and the efficiency
of the latest techniques for automatic text classification: Support Vector Machine
(SVM), Long Short Term Memory (LSTM), Convolutional Neural Networks (CNN)
and the combination of them CNN-LSTM as well as evaluating their relevance
using different pre-processing strategies such as: stop-words removal, stemming, N-
grams and different weighting schemes (TF- IDF, word embedding) to automatically
analyze tweets written in Moroccan Arabic.
The remainder of the paper is structured as follows. Section 2 exposes the
promising fields that benefit from sentiment analysis as well as the works that
have already performed it. Section 3, describes our methodology in building the
tweets corpus. Section 4, presents the used methods and experimentations. Section 5,
discusses the obtained results and finally, main conclusions are drawn and potential
horizons are highlighted.
Sentiment Analysis of Moroccan Tweets Using Text Mining 599
– Marketing: Actually, it is agreed that social media platforms are a wide environ-
ment for advertising products. To do so, companies seek to target our emotions
not our rationality. Therefore, using sentiment analysis besides the techniques
they are already using would be of great benefit. They will be able to personalize
their offers so that they suit well their customers and as a result, the development
of marketing will be easy with sentiment analysis.
– Politics: Acquiring knowledge of social media sentiments is an efficient method
that could be useful for politicians and political organizations. Nowadays, the
feedback of a population on social media is used by leaders in order to be aware
of any potential menaces or conflicts that might arise within their organizations.
– Finance: Thanks to sentiment analysis, investors can spot the promising oppor-
tunities and avoid gambling. They are able to move easily behind their preferable
companies with the intention of guiding their sentiment data in real-time.
In the last few years, considerable attention has been paid to mining sentiment
and opinions from text, due to the available amount of documents and messages
expressing personal opinions on the Internet [1]. Shoukry et al. in [4] Compiled
1,000 tweets about several news subjects. They built two lists: the first contained
negative sentiment words, while the second comprised positive ones. They applied
the unigrams and bigrams features and extracted all Arabic stop words in the training
datasets. The SVM algorithm was applied in their experiment using Weka Suite soft-
ware for the classification process. The accuracy of the SVM classifier was 78.8%.
However, their dataset was relatively of small size.
The work reported by [5] compared the performances of machine- learning and
lexicon-based approaches for the Arabic sentiment analysis. They used maximum
entropy, k-nearest neighbors (KNN), Naive Bayes, and support vector machine
(SVM), for the lexicon-based approach, they constructed a dictionary of Arabic
subjective words, and Their MaxEnt classifier achieved an accuracy of 63% while
for the SVM one, it was of 53.76%. The best accuracy (63.58%) was achieved with
KNN.
Almas et al. (2007) [2] proposed a particular methodology of sentiment analysis
for classifying Arabic and English languages. They used specific features extraction
components that proper for the Arabic language. The resulted accuracy was good
but their training set was limited only to extremist words.
Another work worth mentioning is the project done Kumar et al. (2019) [6]. They
proposed a Bidirectional Soft Attention-Based LSTM (sAtt-BLSTM) approach with
600 M. Garouani et al.
CNN for Sarcasm Detection. The approach was investigated using the SemEval-2015
balanced benchmark and an unbalanced one (consist of 15000 sarcastic and 25000
non-sarcastic ones). They achieved an accuracy of 98%.
3 Methodology
The proposed system is built upon the NLP concepts. Therefore, it contains all the
necessary tools to handle the linguistic diversity that characterizes the communication
of Moroccan users on social network Twitter. We followed the proposed framework
of sentiment analysis bellow as a general methodology, where the sentiment anal-
ysis process is divided into three main phases mainly: data collection, cleaning and
annotating, data preprocessing, classification and evaluation (Fig. 1).
We have collected over 36,114 Moroccan tweets written using the three stages:
– In the first stage, we used the streaming API that allows getting the tweets in
real-time. We have selected tweets written in Arabic that meet the ‘Morocco’
Sentiment Analysis of Moroccan Tweets Using Text Mining 601
geolocation requirement. We were able to extract 7,578 valid tweets from 23,743
tweets collected between January 21, 2020, and February 24, 2020.
– In the second stage, we used SocialBakers to determine the most active Moroccan
accounts. That gave us a list of 29 names. Then, based on this list, we were able to
extract 5,537 valid tweets from 10,371 tweets collected using the user_timeline
API.
– In the third stage, we resorted to Trendsmap to determine the most popular hash
tags in Morocco, we got about 950 distinct hash tags. This collected data are then
used to download tweets using the search API. We were able to extract 2,345 valid
tweets after filtering the Arabic tweets.
A. Final Corpus
Our final corpus consists of Moroccan tweets that were collected between January
21st and February 24th, 2020. Among the 36,114 collected tweets, the 13550 that
were selected as valid tweets were divided as follows: 8750 were published by men
602 M. Garouani et al.
while 4100 by women and for 700 tweets, the gender of the authors could not be
identified because of the lack of information (Table 1).
B. Data Cleaning
A major issue that faces us when we were dealing with the text posted on Twitter
is the unstructured, noisy, and the informal nature of the posts due to the spelling
mistakes, nonstandard words, lack of standardization, and repetitions. The majority
of the collected tweets are written informally, do not follow any grammatical rule,
and contain many abbreviations and errors. Hence, it is essential to undergo the
preprocessing process to minimize the effect of this informality on our classification.
To improve the accuracy of our classifiers, we detected the misspelled words and we
corrected them. Below, we describe some of the most common methods used during
the preprocessing phase (Table 2):
• Delete numbers, punctuation, web links, emails address, emoticons, Hashtags.
• Remove diacritics (harakat, )
• Eliminate duplicated characters that users use to affirm the meaning ( )
C. Dataset Annotation
Until now, there are no annotated linguistic resources for tweets written by Arabic
Dialect speakers (Moroccan, Tunisian, Algerian…). Remember that each country
has its different dialect. In this work, we focused on tweets written by Moroccan
users. Due to the lack of annotated resources, we decided to build our corpus.
D. Statistics
The dataset has four classes of sentiments, Positive, Negative, Neutral and Mixed
(positive + negative), and two language classes used MD and SA. The distribution
of the data according to their classes is illustrated in the following figures (Fig. 2).
4 Dataset Experiments
The classifiers: SVM, CNN, LSTM, CNN-LSTM, NB and LR form the algorithms
we have chosen to carry out this study. To train the algorithms, a 3-fold stratified
cross-validation method was used. For the features representation, we used the TF-
IDF (Term Frequency-Inverse Document Frequency) and word embedding represen-
tations based on the n-grams features (unigrams, bigrams and trigrams). Unigrams
are the simplest features extraction method, while bigrams and trigrams provide the
ability to capture negation or sentiment expression patterns [7].
One of the main purposes of this work is to investigate how preprocessing the tweets
written in Moroccan Arabic (standard or dialect) could enhance the results of the
sentiment analysis of these tweets.
5 Dataset Experiments
Tables 6 and 7 shows the different scenarios settings that we have evaluated. The first
column represents the used classifier name. The second column indicates the used
features and whether a Stopwords filter was used (1) or not (0). The third column
shows the accuracy of each corpus used (only SA, only MD and the fusion of the two
corpora). In the fourth column, we have trained the classifiers in the whole corpora
and we carried out separate tests (a test on SA corpus and MD corpus).
The conclusions drawn from the experimental evaluation results of the combina-
tion of classical classifiers with a TF-IDF weighting system presented in Table 6 can
be summarized as follows:
(1) The SVM classifier using the different grams (Unigrams and Bigrams) achieved
a very good accuracy (84.75%) in the first corpus SA using bigrams without
eliminating stop words, and an accuracy of 80.05% in the third corpus MSDA.
(2) Similarly, the Logistic Regression classifier achieved 82.23% of accuracy for
the corpus SA, using the unigrams without eliminating the stops words, and
78.55% for the corpus (MSDA) by eliminating the Stop Words.
Concerning Table 7, the results of the combination of the Deep Learning classifiers
with the Word Embedding representation model were presented as follows:
(1) The CNN classifier achieved accuracy results of 91.78% keeping the stop
words, and 91.39% eliminating them in the first corpus (SA), and attractive
results in the dialectal corpus with an accuracy of 85.37%, and 89.56% in the
combination of the two corpora.
(2) The combination of the two classifiers CNN- LSTM without elimination of
the stop words obtained a good result with an accuracy of 91.83% in the first
corpus (SA), and 85.55% in the corpus MD.
Task 1:
Table 8 presents the results of classification according to the accuracy for the second
task. The best results achieved in term of the accuracy metric are 88.86% using
LSTM without elimination of Stopwords and 87.95% by filtering them.
From Tables 8 and 9, we can make the following observations:
1. In the experiments, removing stop words decreased all the classifiers perfor-
mance by almost 0.6%, which showed their good effect on the semantic meaning
of the text.
2. In general, the Bigrams function and the inclusion of Stopwords have obtained
the best accuracy results in all classifiers.
3. According to the results obtained, it is clear that deep learning algorithms
(LSTM, CNN, CNN-LSTM) outperform other classical algorithms such as LR
and SVM.
The methods, size of corpora and performances of some state-of-the art papers
are summarized in Table 9.
Table 8 Results of
Model Features SW filter Accuracy
Identification of the used
language LSTM Word embeddings 0 88.86
1 87.95
CNN-LSTM Word embeddings 0 88.43
1 87.06
SVM TF-IDF 0 88.39
1 87.25
LR TF-IDF 0 88.17
1 87.02
Sentiment Analysis of Moroccan Tweets Using Text Mining 607
6 Conclusion
References
1. Pang B, Lee L (2008) Opinion mining and sentiment analysis. Found Trends Inf Retrieval
2(1–2):1–135
2. Yousif A, Ahmad K (2007) A note on extracting sentiments in financial news in English, Arabic
& Urdu. In: The Second Workshop on Computation, al Approaches to Arabic Script-Based
Languages, pp 21–22
3. Mohamed E, Elfeky M (2010) Mining Arabic business reviews. In: IEEE International
Conference on Data Mining Workshops (ICDMW), 13 December 2010, pp 1108–1113
4. Amira S, Rafea A (2012) Preprocessing Egyptian dialect tweets for sentiment mining. In: The
4th Workshop on Computational Approaches to Arabic, AMTA 2012
5. El-Halees A (2011) Arabic opinion mining using combined classification approach. In:
Proceeding of the International Arab Conference on Information Technology
6. Kumar A, Sangwan SR, Arora A, Nayyar A, Abdel-Basset M (2019) Sarcasm detection using
soft attention-based bidirectional long short-term memory model with convolution network.
IEEE Access 7:23319–23328
7. Khreisata L (2009) A machine learning approach for Arabic text classification using N-gram
frequency statistics. J Inform 3(1):72–77
608 M. Garouani et al.
8. Heikal M, Torki M, El-Makky N (2018) Sentiment analysis of Arabic tweets using deep
learning. Procedia Comput Sci 142:114–122
9. Al-Twairesh N, Al-Khalifa H, Al-Salman A, Al-Ohali Y (2017) AraSenTi-tweet: a corpus for
arabic sentiment analysis of Saudi tweets. Procedia Comput Sci 117:63–72
10. Nabil M, Aly M, Atiya A (September 2015) ASTD: Arabic sentiment tweets dataset, pp 2515–
2519
11. Nawaf AA, et al (2015) Towards improving the Lexicon-based approach for Arabic sentiment
analysis. Int J Inf Technol Web Eng
Contribution of Digital Collaboration
and E-Learning to the Implementation
of Smart Mobility in Morocco
Abstract This article attempts to highlight the need to rethink, more intelligently in
Morocco, how to manage mobility, a major pillar of economic competitiveness and
social well-being, by first initiating a broad reflection on the very needs of displace-
ment, using as much as possible digital collaboration, e-learning and distance training
and then the massive use of Intelligent Transport Systems in order to ensure optimal
exploitation of the existing infrastructure before to launch cumbersome investment
for irreversible infrastructure projects. Our analysis has shown that with the compre-
hensive diagnoses made and the clear and consistent strategies that have been adopted
in Morocco, there is only a tiny link missing to put digital and new technologies to the
benefit of human and sustainable development. Indeed, it only takes a national and
unified mobilization to move from the stage of digital emergence, to the transition
and then to the transformational phase. A digital ecosystem is emerging, bringing
innovative solutions through artificial intelligence, virtual reality, and augmented
reality, which need to be deployed to measure their reach and their limits and try to
evolve them to meet increasingly diverse and constraining collaboration and learning
needs.
1 Introduction
Digital technologies around the world are bringing structural changes in industries,
economies and social relationships. The agility they offer allows for cost optimization
and the resolution of the most complex problems.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 609
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_55
610 M. M. Maaroufi et al.
Since the approval of the new constitution in 2011, Morocco has taken steps to
achieve a digital transformation in the service of the administration, the economy
and its cities.
The Royal speech, at the opening of the legislature on 13/10/2016, outlined the
strategic directions inviting the executive to use new technologies for an upgrade of
the administration in favor of the citizen in Morocco. His Majesty the King said on
October 13, 2017 that Moroccans will need a quality education that allows them easy
access to the labor market.
The king’s new development model will be a boon for a shared, integrated and
comprehensive vision for the next ten or fifteen years and will provide a framework
where humans are at the center of development.
It is therefore obvious that in order to make Morocco a country of solidarity and
prosperity, we must avoid the urban/rural digital divide by opting for smart cities but
also for an attractive and connected rural world [1].
After describing the problem, we will first review the initiatives and strategies
concerning mobility, training and digital in Morocco and in order to encourage inno-
vation within Moroccan digital startups, we will then present a prototype virtual
platform developed by a Moroccan startup that can answer many questions using
Virtual Reality and Increased Reality. Finally, we will conclude with a few recom-
mendations. At no time do we claim the completeness of the findings, analyses or
solutions, these are only strands that can develop further and continuously.
2 Research Problem
Smart cities can be identified by six closely linked levers: smart economy, smart
mobility, smart environment, smart people, smart lifestyle and a smart administra-
tion [2]. So, we can’t call it a “smart city” without talking about “smart mobility.”
Indeed, the 4th industrial revolution and the digital transition have led to a rethinking
of how to manage mobility: Designing and building road infrastructure and imag-
ining new services for “connected mobility” without excessive waiting or congestion
at interconnections. Provide agile, functional and cost-effective solutions for opera-
tions and maintenance (smart sensors, user information solutions, infrastructure and
intelligent transport systems, etc.).
While the term “smart” refers to better use of technology, intelligent mobility is
also defined as “a set of coordinated actions to improve the efficiency and environ-
mental sustainability of cities. [3].” The term “mobility” highlights the preponderance
of humans in relation to infrastructure and vehicles.
In the face of traffic congestion, the solutions currently adopted, which consist of
addressing the problem by developing and expanding infrastructures, only encourage
increased travel and postpone the problem with more criticality. It is true that beyond
a certain density of traffic, even ITS, are out of date. Hence the need to act on
the problem at the root level by optimizing the very need for displacement. Such
action can only be successful through the provision of innovative solutions to serve,
Contribution of Digital Collaboration and E-Learning … 611
collaborate and train remotely. At this point, dynamic “mobility” technologies will
be able to manage and smooth traffic efficiency. It is only by combining agility,
efficiency and sustainability that we can speak of “smart mobility”.
Therefore, the questions that need to be asked are: what “smart mobility” to adopt
in Moroccan cities? Is it just “connected mobility”? Or, first, a “sustainable mobility”
that contributes to a humane, inclusive and resilient city and which is, above all, a
thoughtful, optimal and coherent mobility? And how can digital collaboration and
distance learning help build “smart mobility”?
100 metropolises have committed to the “Mobilize Your City (MYC)” initiative
aimed at a proactive transition and a profound cultural and organizational transfor-
mation of people’s movements and related logistics activities in order to make cities
more inclusive and reduce their Green House Gas (GHG) emissions. This initiative
builds on the Networks’ Collective Intelligence and existing planning practices and
incorporates the principles of integration, participation and evaluation.
Morocco is one of the first countries to be engaged in the MYC initiative at COP22
in Marrakech (2016). The desire to include mobility planning in 26 municipalities, in
the Sustainable Development approach, resulting from the MYC project, was aimed
at:
– Improving the quality of life and economic attractiveness of the city;
– Offer alternatives to individual vehicles;
– Put in place a Sustainable Urban Mobility Plan (SUMP).
Morocco has seen significant progress and reforms in the areas of the environment,
Sustainable Development (SD) and the fight against Climate Change (CC) over the
past decade. Several sectoral strategies, including transport and logistics, integrate
the environmental dimension. The economic stakes are important since the cost of
pollution in the area in Morocco represents more than 10 Md DH (1% of GDP) [4].
The National Sustainable Development Strategy (SNDD 2030) identified the
transport sector as the third largest energy consumer in Morocco in 2014. It accounts
1 https://mobiliseyourcity.net/.
612 M. M. Maaroufi et al.
for 16% of total emissions and 28% of emissions from the energy module. Sustain-
able mobility is defined as “a transport policy that seeks to reconcile accessibility,
economic progress and the reduction of the environmental impacts of selected
transport” [5].
In addition, Morocco is the first country to have undertaken an adaptation of the
global macro-roadmap for the transformation of transport from the Paris Process on
Mobility and Climate (PPMC). Morocco’s 2018 roadmap, which aims to support
national strategies and in particular the SNDD, recommends less travel for more
productivity and better quality of life. Indeed, time spent in transport limits the
productivity of the economy and the quality of life of citizens. Less travel means more
time and resources for value-added activities and social life. The roadmap also calls
on governments to move forward with dialogue between employers’ and employees’
organizations in favor of remote work in order to reduce commuter kilometers.
SNFP 2021 aims for quality vocational training everywhere, for all and throughout
life, at the service of the development and enhancement of human capital and for a
better competitiveness of the company [6]. This strategic vision is oriented in three
directions: The citizen, the enterprise and the state/territorial communities.
In order to have a flexible, responsive and professionally rooted vocational training
system, SNFP 2021 has five strategic objectives:
– Ensuring the right to vocational training: social inclusion and territorial extent
through the training of 10 million citizens;
– Improving the company’s competitiveness as an actor and a privileged training
space: Training, from 2021, of 20% per year of employees and 50% of trainees
in professional settings;
– Maximizing professional integration through continuous improvement in the
quality of training: 75% of the winners’ integration rate by 2021;
– Have on-demand training and ensure a better attractiveness of vocational training
to enable young people to express their vocations;
– Renovate and strengthen the governance of public policy of vocational training
inducing performance and synergy of the system.
As part of the ‘Digital Inclusion Vector Distance Training’ project, the ‘Devel-
opment of Distance Training as a New Training’ project aims to dedicate and insti-
tutionalize distance learning as a form of training and graduation and certification
in addition to strengthening digital inclusion. The specific objectives of this project
are:
– Improving access to vocational training (young people, workers, people with
special needs, housewives, inmates, etc.)
– Optimize human and financial resources dedicated to vocational training.
Contribution of Digital Collaboration and E-Learning … 613
The strategic vision for the 2015–2030 reform, which has been translated into
Framework Act 51.17 on the education, training and scientific research system,
aims to establish a new school with three main foundations: equity and equality
of opportunity, quality for all and the promotion of the individual and society [7].
From this perspective, the first objective of change is the transition from the logic
of the linear transmission of knowledge and memorization to a logic of learning, the
development of a sense of criticism, the construction of a personal project, the acqui-
sition of languages, knowledge and skills, values and digital technologies. Thus, the
new school will promote the transition from a knowledge-consumption society to a
society that produces and disseminates it, through the mastery of digital technolo-
gies and the development of scientific research and the culture of innovation and
excellence.
Since 1996, Morocco has been part of the restructuring of the field of informa-
tion technology by the adoption of Law 24–96, which has enabled the launch of
a first step in the liberalization of the telecommunications sector. In this context,
Morocco has adopted multiple digital strategies: the 1999–2003 five-year plan, e-
Morocco 2010 covering the period 2005–2010 and the MN2013 strategy, which is
being implemented over the period 2009–2013 [1].
In order to further boost the development of the digital economy sector, a new
vision for digital technology by 2020 has been developed. The first of the three pillars
of the strategy aims to transform the national economy digitally through the reduction
of the digital divide for the benefit of citizens (connectivity for all, primary, secondary
and higher education programs, public access to WIFI) and businesses in particular
small and medium companies/very small companies (regions disadvantaged in terms
of connectivity).
The Agency of Digital Development (ADD), created in 2017 with the main
mission of implementing the state’s digital development strategy, has drawn up a
note in 2019 to propose to the government guidelines for the development of digital
by 2025.
In December 2019, the prospects for digital development in Morocco were defined
with four main challenges:
a. A digital administration (Smart Government) at the service of citizens and busi-
nesses in order to reduce the rate of dissatisfaction of citizens/companies from
2 https://www.add.gov.ma/.
614 M. M. Maaroufi et al.
85–90% to less than 20% through the transformation of interactions with public
administration through end-to-end digitalization of citizens/business paths;
b. A competitive economy thanks to the performance gains brought by a digital
ecosystem and innovation through the Start-up Program, which consists of the
installation of a fabric of more than 2,500 startups, over the next five years,
Digital park, Smart factory, Digital SME: a complete device for digitalization
of Small and Medium-sized Enterprises (SMEs), and Digital self-entrepreneur.
Thus, innovation and digital can be a lever that creates wealth;
c. An inclusive society through digital technology. Digitalization can open up
new opportunities by giving less advantaged populations access to information,
social benefits, care services, and education. It offers, for example, digital classes
in Moroccan schools located in remote areas.
d. An environment and digital trust. This component mainly concerns the estab-
lishment of a regulatory framework, a national programme of higher educa-
tion, vocational training and continuing training for digital professions, called
‘Digital Generation’ and its support ‘Digital Infrastructure’ for the establish-
ment of an advanced infrastructure (very high-speed, 5G, government cloud,
regional data centers…) necessary for the digital transformation of the country.
It is also planned to contribute to the launch of an ‘AI Ecosystem’ program on
artificial intelligence thematic research, based on cases of sectoral use, with the aim
of coordinating around the targeted research topic the prime contractors or end-
customers, innovative startups and public and private training bodies. This project
aims to set up an ecosystem dedicated to artificial intelligence on the natural language
processing and natural language generation components for digital inclusion with a
conversational AI engine.
With 4.3 million vehicles, the car fleet more than doubled in Morocco between
2002 and 2018 [8]. Travel trends and needs are also increasing, leading to increased
demand for people and goods.
The management of mobility in cities, undergoing an acceleration of the
phenomenon of urbanization and rural exodus, is increasingly becoming a thorny
constraint. Failure to do so will inevitably prevent their transformation into smart
cities.
Road transport infrastructure is under pressure. The solutions adopted to date
are always on infrastructure with large-scale, extremely expensive and definitive
actions (cable-stayed bridges, hoppers, road widening, etc.). Improved infrastruc-
ture generates additional induced traffic and existing diverted traffic, causing further
Contribution of Digital Collaboration and E-Learning … 615
With regard to e-learning, SNFP 2021 has only planned the design and development
of an e-Learning platform [6]. Virtual Reality and Augmented Reality based on
Artificial Intelligence, essential for the practical assimilation of several disciplines,
have not been programmed. At the same time, dealing with near-reality situations
through augmented reality can also energize face-to-face training in an innovative
Blended Learning approach.
Indeed, transmitting knowledge to learners through Immersive Learning is the best
way to enable practice and move into Agile Mode with Virtual Training, experiential
learning, immersive simulations and digital training modules.
SNFP 2021 has planned several actions scheduled for 2017–2018 in order to offer
a demand-driven training offer. However, it has not put in place a process for the
retraining of young graduates, self-entrepreneurs and workers who will be brought
into line with the new tools and concepts of the labor market.
The SWOT analysis, conducted in this work, summarizes the diagnostic elements
of the situation of digital transformation in Morocco. It sees the strengths of the
digital ecosystem, consolidated through the various strategies developed and raises
the shortcomings, which remain to be perfected. This analysis is also an opportunity
to recall the opportunities that can boost the implementation of digital in the Kingdom
and outlines some risks that may hinder its implementation (see Table 1).
616 M. M. Maaroufi et al.
Provided in Platform as a Service mode (PaaS), this digital solution is accessible from
any connected device. It is fully customizable and adapts to any type of event or exhi-
bition, public or private service, training or conference, meeting or networking space
or open houses. This virtual meeting place, with innovative features (live streaming,
machine and instant translation, Artificial Intelligence, Smart Data), offers an acces-
sible platform, dematerialized logistics and online management tools for admin-
istrations, universities, public facilities, marketplace-B2B, museum, media library,
conference and seminar center, exhibition halls, showroom.
A true 3D virtual replica of the “hard” environment, only the tools change! 3D
customization allows you to choose the layout and layout of virtual spaces (walls,
floor, visual and display colors, animation videos) to the identity of the exhibitor or
speaker (see Sect. 5.4 Models of the prototype).
No more stress of late printing, orders for furniture leading to chaotic deliveries. The
brochures, and training materials become downloadable files and the virtual screen
allows you to view all presentations.
The virtual platform is a matching space, matching a dynamic offering to a diverse
and unpredictable demand via an intelligent preference system that adapts and capi-
talizes to submit categorized and qualified suggestions. For example, reception areas
in the form of information desks, thematic halls with virtual booths for a permanent
or one-off exhibition of products and services, conference rooms and workshops,
and training rooms at Augmented Reality allow for an ideal responsiveness in Agile
Mode. The platform includes:
– An essential complement and an extension of physical events;
618 M. M. Maaroufi et al.
– A true Eco-responsible initiative preserving the environment (less travel, less air
pollution and noise pollution, less waste and materials used to build hard spaces);
– A time-saving opportunity and a contribution to road safety;
– Unlimited accessibility space, including People with Reduced Mobility;
– A capitalization assured through 100% rise;
– Insurance of no risk of cancellation in the event of strikes, epidemics or natural
disasters;
– An alternative to making automated webinars with the possibility of ques-
tion/answer interactivity;
– Immersive and intuitive navigation in real time.
The virtual platform, developed by a Moroccan startup, was able to succeed several
events in different editions. Thus, the 2nd edition of the Virtual Building and Public
Works Show BTPEXPO3 in 2016, was the first virtual exhibition in the world to
receive the COP22 label and saw the participation of 200 national and international
exhibitors and more than 20,000 qualified visitors, as well as the broadcast of 9
thematic conferences over 3 days.
On the occasion of the fiftieth anniversary of its creation, the National Docu-
mentation Centre organized in 2016, using this platform, the first Moroccan Virtual
Interactive Show of the standby and the documentation DOCEXPO4 .
Other successful events have been organized such as the Virtual Forum of Self-
Entrepreneurs AEXPO5 and the Virtual Immovable Show IMMOEXPO6 .
It appears that digital collaboration and e-learning is not yet widely used in Morocco.
The integration of Virtual Reality and Augmented Reality as in the prototype
presented in this article will be able to solve certain problems related, for example to
practical work in teaching and the non-friendly and static environment in traditional
collaboration and service platforms.
The scope of possibilities for remote work offered by virtual reality is consid-
erable. Indeed, new ways of working together from a distance are emerging. More
immersive, more realistic and replicating our professional environment. But the issue
of telework is broader than that because working at home requires skills such as
3 https://www.youtube.com/watch?v=kRv7LdWx5-0.
4 https://www.youtube.com/watch?v=IDoPO9oAJL0.
5 https://www.youtube.com/watch?v=V0kDckSk7ko.
6 https://youtu.be/7S0h2NQfJ8M.
Contribution of Digital Collaboration and E-Learning … 619
References
1. CC (Court of Auditors) (2014) Summary of the evaluation of the Morocco Digital Strategy 2013,
13 p
2. Giffinger R, Fertner C, Kramar H, Kalasek R, Pichler-Milanovic N, Meijers E (2007) Smart
Cities: Ranking of European Medium-Sized Cities. Centre of Regional Science, Vienna, 12 p
3. Benevolo C, Dameri RP, D’Auria B (2016) Smart mobility in smart city. In: Torre T, Braccini
A, Spinelli R (eds) Empowering Organizations. Reading Notes in Information Systems and
Organisation, pp 13–28
4. METLE (Ministry of Equipment, Transport, Logistics and Water) (2018) Roadmap for
Sustainable Mobility in Morocco, 19 p
5. SEDD (Secretary of State to the Minister of Energy, Mines and Sustainable Development,
responsible for Sustainable Development) (2017) National Sustainable Development Strategy
2030. Executive Summary, 60 p
6. SEFP (State Secretariat for Vocational Training) (2015) National Professional Training Strategy
2021, 89 p
7. CSEFRS (Higher Council for Education, Training and Scientific Research) (2014) Strategic
vision for the 2015–2030 reform, 27 p
8. METLE (Ministry of Equipment, Transport, Logistics and Water) (2019) The METLE in
numbers 2018, 37 p
MOOCs Semantic Interoperability:
Towards Unified and Pedagogically
Enriched Model for Building a Linked
Data Repository
1 Introduction
At the end of 2019 and during 2020, the spread of COVID-19 caused the shutdown of
all educational institutions around the world. In order to ensure the continuity of the
learning process, many universities have turned to E-learning and especially MOOCs.
The number of users of MOOCs has doubled compared to 2019 [1] before the spread
of the virus. Many instructors adopt MOOCs to provide their academic courses.
However, there are challenges to overcome, such as the design of the course. This
work is part of a general framework of recommendation system [2], which plans to
benefit from the existing models, reduce the workload and limit the expense of making
MOOCs. The aim here is to help the MOOC makers find suitable and personalized
content from different providers. However, each provider handles different formats
or schemas of metadata or vocabulary; this infers an issue of heterogeneity.
Our purpose is to gather those different pieces of MOOCs in a new MOOC, and
this in spite of the basic heterogeneity between resources. This is the premise of the
idea of interoperability. The interoperability of educational resources has been the
subject of several studies, both at the standardization level, and at the architectural
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 621
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_56
622 H. Sebbaq and N. E. Faddouli
the metadata of LOM and Dublin Core, as it seeks to allow both the aspects of LOM’s
“learning object” and the Abstract Dublin Core Model [8]. For this purpose, MLR
can be the solution to the interoperability problem between educational resources.
In 2013, the Dublin Core initiative and several major players in the IT industry,
including Google, Bing and Yahoo search engines, support the development of the
LRMI standard [9]. Their primary purpose is to incorporate elements of educational
metadata. Based on these elements, the LRMI will extend the interoperability to the
worldwide search engine level and will be able to allow specialized research [9].
Semantic interoperability is one of the challenges facing the world of E-learning
at present. Different learning repositories freely exchange a large number of online
content and data, and these repositories are based on various implementation
concepts, making the integration process expensive. Our research work aims to
integrate existing MOOCS providers and is focused on current learning systems
comprising very rich and varied content. Standards such as LRMI could be an
adequate solution, but because it is only recommended for systems under develop-
ment, it cannot be adopted. Our work therefore focuses on the integration of MOOCs
in particular, while LOM and Dublin Core do not cover the MOOC field. The LRMI
integrates such MOOC specific fields, including the type of learning resource and the
time needed, but does not cover other areas such as learning objectives or pedagogical
model necessary for our work.
The above restrictions push us to propose a unified and specific model for MOOCs.
Our aim is to ensure semantic interoperability between the various MOOC providers
and to provide a shared understanding of the terms that will be exchanged. In order to
address the problems of semantic heterogeneity and interoperability of MOOCs data,
semantic web technologies based on ontologies and linked data appear as a solution.
The strength of ontology lies in its ability to describe information closely connected
by syntactic or semantic relationships [11]. Those ontologies are used by Linked
Data to link and integrate data from different sources and with different formats. In
the field of education, linked data is relatively unexplored; we have already checked
this in our previous work [2].
Linked Data Repository. The authors of [18] provide a semantic Web-based solu-
tion with an ontology of concepts and pedagogical skills, a repository of semantic
educational resource annotations, and a Web service based on SPARQL. They high-
light the implementation of the services provided by the current solution and others.
Their approach covers both phases: the proposal for ontology and the creation of a
Linked Data repository. However, it does not cover MOOCs. Similarly, in order to
meet the needs of learners or teachers belonging to various universities, the authors of
[19] propose a distributed warehouse model that enables the indexing of educational
objects based on the principle of ontology. They propose an architecture, based on
a classic relational database, of a distributed educational warehouse. They are inter-
ested in university-level educational services (digital or not). This approach is costly
because of its focus on a process of indexing and semantic annotation at each univer-
sity’s level. Likewise, researchers in [20] provide a data repository of books based
on LOM, Dublin Core and IMS, which are cut up and then annotated. For their part,
authors in the paper [21] propose a model of data repository for OERs but without
ontology or Linked Data being used. Similarly, the authors of the paper [22] do not
take use of ontology or Linked data to offer their data warehouses. Their work is about
classifying MOOCs by directly attacking the data extracted through web scraping
according to the learning style. In [23], a Data Warehouse is also suggested. The
automatic generation of SPARQL Build queries is used, with unified target vocabu-
lary terms defined in the Build clause, which are then sent to run on various SPARQL
endpoint sources. RDF graphs with a single vocabulary are the retrieved data.
There are several limitations to the works above. First, some of them do not cover
the MOOC domain. Second, while some propose a MOOC ontology, they do not
MOOCs Semantic Interoperability … 625
cover the concepts necessary for our research, such as pedagogical model or learning
objectives. Third, most of them do not go as far as proposing an architecture for a
data repository. Fourth, in setting up the data repository, they do not rely on Linked
Data.
We propose an ontology designed to capture the broad scope of MOOCs and unify
their representations. This ontology will allow us to build a MOOC metadata repos-
itory based on linked data principles (see Fig. 1). This repository will serve as the
basis for our recommendation system [2] and may be used by other researches.
The contribution of this approach lies on three folds. First, it relies on the Bloom’s
Taxonomy to propose an ontology enriched with metadata for the characterization
of the pedagogical dimension; second, our ontology will serve as a standard to unify
the representation of MOOCS and can facilitate interoperability between MOOCs
platforms. Third, it provides an automated collection of metadata from MOOCs and
builds a repository accessible through a SPARQL endpoint that can be used for other
purposes.
We start with the search for a theoretical framework that will serve as the basis
for our approach. Later, we propose an ontology designed to capture the wide scope
of MOOCs and to unify their representations. In order to clarify and concretize the
epistemological position of the pedagogical dimension of MOOCs, we perform a
mapping between the learning objectives and our theoretical framework.
above guided the research of Conole and Magrayan in the development of Evalua-
tion Frameworks. This evaluation is based on open questions and often requires the
intervention of an expert. For their part, [29] adopts the “web-based online instruc-
tion” theory, which is more general and integrates three global dimensions of design
elements: information, instruction and learning.
The aim of our research is the large-scale evaluation and classification of the
pedagogical dimension of MOOC on the basis of its learning objectives. Bloom’s
Taxonomy is the most appropriate one for our context since it covers the different
levels of cognitive learning and allows classifying the learning objectives on its six
hierarchical levels [31]. The initial purpose of Bloom’s taxonomy is to assist teachers
in the development of rubrics and in measuring achievement of their learning goals
by providing guidelines [30, 31]. The research paper [31] proposes a revision of
the original taxonomy. It defines a two-dimensional framework: Knowledge and
cognitive processes. The first dimension takes the subcategories of the first level of
the original taxonomy; the second dimension takes the six levels and renames them
in verb: remembering, understanding, applying, analyzing, evaluating, and creating.
For our research, we take into account the dimension of the cognitive learning
process but we reduce it to three levels. We used a modification of Bloom’s Taxonomy
adapted from [32]. The following Table 1 gives a summary of this adaptation.
MOOCs Semantic Interoperability … 627
4 Implementation
The first module of the architecture of our Linked Data Repository is “Ontology
Design and Modeling (see Fig. 1). This module consists of the modeling and devel-
opment of a semantic ontology. The two following modules are “Data Extraction”
and “Data Integration”. The first module seeks to extract raw data from MOOC
providers, while the second one aims to convert and integrate this raw data according
to our ontology. Finally, we use the Apache Jena Fuseki Server to build and set up a
SPARQL Endpoint, and execute SPARQL queries on the created RDF data.
The data is collected from the two providers edX and Coursera. Coursera provides a
RESTful API for accessing its catalog via HTTP GET. This API exposes all Coursera
courses. It is publicly available and does not require authentication. The data is
retrieved in JSON format. Coursera also provides various optional parameters that
can be included in the API request. For its part, edX does not provide an API for
accessing its catalog. To extract data from edX, we opt for Web Scraping. Several
types of libraries and Framework allow us to do Web Scraping. In particular Scrapy
[34], Selenium [35], BeautifulSoup [36], to name only the most famous. For our
work, we use Scrapy. Scrapy is an open source and collaborative framework for
performing requests, scraping and saving data on the web, it is sufficient on its own
to build a web scraping project. Scrapy’s functionality can be extended by importing
other libraries. For our project, we use also splash [37] for the extraction of pages,
which contains JavaScript, and selenium for navigation between web pages.
The “Data Integration” module takes JSON files as input and converts them to an
equivalent Java object using the “gson” library. We use after the JENA Framework
[38] to perform an RDF mapping.
After we convert the raw data to the appropriate RDF format, the data resides in
a stable location where SPARQL queries [39] will be made. For this purpose we
use the Apache Jena Fuseki server which is an RDF store as well as a SPARQL.
There are several modes in which the Apache Fuseki server can operate, either as an
operating system service, as a Java web application, or even as a stand-alone server
[40]. Apache Jena Fuseki Server provides an intuitive graphical user interface in
which the user can upload multiple files to which SPARQL queries are made (see
Fig. 3).
MOOCs Semantic Interoperability … 629
5 Conclusion
In this research paper, we tackle the problem of the interoperability of various educa-
tional resource repositories, especially in the domain of MOOCs. To resolve this issue
we propose, in one hand, an ontology that unifies the representation of MOOCs and
allows the interoperability between the different providers despite the heterogeneity
of their data schema. On the other hand, we implement a Linked Data repository
based on the ontology constructed as well as the semantic web technologies. This
repository will serve as the basis for our recommendation framework.
In order to concretize the epistemological position of the pedagogical dimension
of MOOC, we carried out a mapping between MOOC’s Learning Objectives and
Bloom’s Taxonomy levels. However, one of the limitations of our current research
work is that the MOOC classification remains manual, which is tedious given its large
scale aspect. In a future work, our proposal will be the automation of the pedagogical
classification of MOOCs.
References
Abstract The combination of Internet of Things (IoT), Wifi network and other
advanced technologies have enabled universities to transform from classical envi-
ronment into smart, digital and connected universities. Modern universities create
new way for students to learn, simplify operations for university administrators,
provide a safer and a secure environment, improve environmental sustainability, and
more. This research describes smart university as a new concept and focuses on how
IBN TOFAIL University can become a smart environment.
1 Introduction
The emerging technologies like: Internet of Things, Big Data and the Cloud
Computing changed the way we live our lives as the information is always available
anytime and everywhere. Also this advancement has many impacts on the world
around us.
University can be as large as cities. The main idea is to transform a classical
university into a smart university based on the concepts and experiences of smart
cities. In fact, smart university is similar to a smaller version of smart city, it can be
a physical or digital environment in which humans and technologies interact with
each other to enable and create new experiences or services for students, professors,
and employees also to improve operational efficiency.
Recently the number of smart universities is growing faster than expected which
can provide a better academic environment for the students. Around the world,
different universities have earned the ‘smart’ tag. In fact, Stanford University, Curtin
University, Arizona State University and the University of San Francisco are just a
few examples of universities which adopted smart university technologies to remain
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 633
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_57
634 K. Mazwa and T. Mazri
competitive in the higher education landscape [1]. The purpose of this research is to
transform IBN TOFAIL University from a traditional into smart university based on
the smart city concept.
In this regard, this paper is divided to seven sections. In Sect. 2 and 3, we summa-
rized the main concepts of smart city and smart university. Section 4, describes our
vision to transform the University of IBN TOFAIL to a smart university. Section 5,
presents the methodology used in our vision. For Sect. 6, we identified the main
systems could be involved in our smart university. In Sect. 7, we discussed the chal-
lenges and issues that could face our smart university. Finally, the conclusion is in
Sect. 8.
Smart city has emerged as an important concept at the end of the year 2000. It
has gained huge amount of attention from professionals, academics, and researchers
from multiple disciplines. A global phenomenon whose success is indisputable, the
smart city remains enigmatic without precise definition. For Smart city concept,
the researchers propose many definitions and multitude forms for example “Digital
City”. Four important axes are included in those definitions [2]:
• Managing the city’s features by using the Information and Communication
Technologies (ICT).
• Creating for the residents a high quality of life.
• The central role of citizens by making cities based on citizens needs.
• Smart cities develop new efficient services in numerous areas, for example
informatics, urbanization or sociology.
Actually, the smart city and the sustainability are now side by side. Based on the
analysis and according to the International Telecommunication Union Focus Group
on Smart Sustainable Cities, the most suitable definition for a smart sustainable city
is:
“A smart sustainable city is an innovative city that uses information and commu-
nication technologies (ICTs) and other means to improve quality of life, efficiency
of urban operation and services, and competitiveness, while ensuring that it meets
the needs of present and future generations with respect to economic, social and
environmental aspects” [3].
This definition highlights the interdisciplinary areas of the smart cities concept,
which present the city as a set of services to provide better solutions while improving
quality of life.
IBN TOFAIL UNIVERSITY: From Classical University … 635
The University of IBN TOFAIL is a generalist university that was established in 1989.
It includes a multiple institutions and several branches in different areas such as liter-
ature, humanities sciences, law, economics, sciences, education, communications,
business, management, computer science and telecommunications. The university
has a variety of faculties and engineering schools. Its employees exceed 200, 490
academic staff and it has over 33000 students [5].
As part of the preservation of the environment, the university is strongly involved
in environmental issues and among its objectives are to support the learning perfor-
mance of every student and to promote the research and innovation field. Our univer-
sity has already many applications in place that require being equipped through IoT
devices and effective systems that can reach the goals of smart university. Smart
universities are designed by combining different technologies like IoT, RFID…etc.,
in order to deliver teaching techniques and research strategy.
636 K. Mazwa and T. Mazri
In this context, the main goal is to implement Smart City concepts to the univer-
sity in order to transform the University of IBN TOFAIL into a reference in envi-
ronmental sustainability. Three important areas are required to reach this goal (see
Fig. 1): Conservation and construction, sustainability and application of advanced
technologies. These three keys create a smart, sustainable, social, and comfortable
university [6].
5 Methodology
Similar to smart cities, many application categories must be interpreted in our smart
university but the main key of this work is ICTs (Information and Communications
Technologies), without substantial advances in ICT our university will continue to
be inefficient and perform poorly [7].
Through the power of ICTs, applying technology is an inclusive mechanism to
control and manage each and every category with various aspects of action:
• Include the measuring equipments like surveillance cameras, sensors…etc. in
every part of the university.
• Connecting the data with the cloud computing applications by using telecommu-
nication networks and protocols.
• Analyzing and monitoring all the collected data by artificial intelligence and
big-data techniques.
• Usage of actuators which have the task of carrying out physical actions using the
AI algorithms.
Supporting every task of the UIT smart university must be provided with the
application of ICTs.
IBN TOFAIL UNIVERSITY: From Classical University … 637
Smart classroom can provide a better learning and teaching opportunities with the
adoption of new high-end digital technology and smart systems. It can be comprised
with [8, 9]:
ID Cards: Those cards can identify students, professors, administration, staff and
visitors using RFID tags to control and to monitor who is inside the building at all
times.
Smart Lighting: The system can help staffs control and personalize light settings by
using a Smartphone application. Or by using occupancy sensors that detect people’s
presence indoor within a specific area to turn lights on or off automatically to save
energy and improve safety.
Interactive Whiteboards: It can help teachers deliver more effective lectures and
it can optimize learning and engage more students into the education process. Those
smart boards can also record everything written.
Smart Door Locks: IoT devices can be installed on door locks to control who has
access to the room.
Smart Windows: It can offer a high level of control over light by monitoring the
amount of light that enters the classroom.
Cameras: By deploying smart cameras, it can control the security systems that can
track individuals and facial recognition applications.
The smart library system is based on the Internet of Things (IoT) and Radio Frequency
Identification (RFID). The books and other assets from the library are adopted with
tag, RFID can improve the efficiency of tracking and monitoring of the books like
book circulation and management of reservations…etc. [10].
The smart library system can be combined with an Android application that is
created and is installed in all the student and staffs phone. This application can inform
the students about the books available in the library, question papers and other books,
magazine information or the status about the borrowed books [11].
Fire detecting system is needed in UIB to protect lives by identifying a developing fire
automatically to alert the university responsible and fire emergency organizations.
Some of the components used for fire detection systems are [12]:
IBN TOFAIL UNIVERSITY: From Classical University … 639
Flame Sensor: is a sensor which is used to recognize the presence of fire or flame.
It will generate flame signal when fire is caused.
Gas Sensor: is a sensor which will detect different types of gas. They are used to
identify gases like carbon dioxide, carbon monoxide and many more.
A university produces a lot of waste that should be managed properly at regular inter-
vals so that the university can be clean and hygienic. The Dustbins in the smart waste
management system can comprise sensors measuring fill level and a communication
system that transfers this data to the garbage trucks or waste disposal unit with the
location and status of each dustbin. Then, a request is automatically generated for
waste collection [15].
The smart universities in many regards are considered as small cities. In this paper,
we identified many kinds of systems and technologies that can transform UIT from
classical to smart university. In facts, our UIT is in the process of adopting innovative
techniques for creating a smart environment.
With the development of technologies and within a modern society, our smart
university will confront many challenges among them the technical obstacle.
640 K. Mazwa and T. Mazri
Many of the technologies discussed above are not necessarily new but they will
just be used in new ways. However, the security and the privacy are still big issues.
Below, we list some challenges and issues that can face our UIT [16]:
Security Attacks: The data from every connected object and device in our smart
university are vulnerable to security threats such as gossip attack, automated invasion
attack…etc. However, strong security mechanism is needed that help defend from
cyber attacks.
Quality of Service: In order to transfer a huge amount of data for various services,
it is possible that we discover lack of quality of services. Therefore, quality of
service features must be implemented to ensure that even in the critical situations
the applications can still function to provide better services to our smart university.
8 Conclusion
Nowadays several concepts are proposed such as Big Data, IoT (Internet of Things),
and the Cloud Computing. All of them gained greater emphasis and our University
of IBN TOFAIL is a great place where all of these areas could be implemented.
In this paper, we investigated the different systems that could be involved in our
UIT which allows for greater security, more interactivity and better learning. In
addition, we give insights on the challenges that could face our UIT smart university
from becoming reality.
References
1. Nuzzaci A, La Vecchia L (2012) A smart university for a smart city. IJDLDC 3(4):16–32.
https://doi.org/10.4018/jdldc.2012100102
2. Dameri RP, Cocchia A (2013) Smart city and digital city: twenty years of terminology evolution,
p8
3. Ahvenniemi H, Huovila A, Pinto-Seppä I, Airaksinen M (2017) What are the differences
between sustainable and smart cities? Cities 60:234–245. https://doi.org/10.1016/j.cities.2016.
09.009
IBN TOFAIL UNIVERSITY: From Classical University … 641
4. Cook DJ, Das SK (2007) How smart are our environments? An updated look at the state of the
art. Pervasive Mob Comput 3(2):53–73. https://doi.org/10.1016/j.pmcj.2006.12.001
5. www.uit.ac.ma. https://www.uit.ac.ma/. Accessed 15 Nov 2020
6. Min-Allah N, Alrashed S (2020) Smart campus—a sketch. Sustain Cities Soc 59:102231.
https://doi.org/10.1016/j.scs.2020.102231
7. ur Rahman M, Himanshi, Deep V, Rahman S (2016) ICT and internet of things for creating smart
learning environment for students at education institutes in India. In: 2016 6th international
conference - cloud system and big data engineering (confluence), Noida, India, January 2016,
pp 701–704. https://doi.org/10.1109/CONFLUENCE.2016.7508209
8. Temkar R, Gupte M, Kalgaonkar S (2016) Internet of things for smart classrooms. Int Res J
Eng Technol 3(7):203–207
9. Bautista G, Borges F (2013) Smart classrooms: innovation in formal learning spaces to
transform learning experiences. Bull IEEE Tech Committee Learn Technol 15(3):18–21
10. Brian ALA, Arockiam L, Malarchelvi PDSK (2014) An IoT based secured smart library
system with NFC based book tracking. Int J Emerg Technol Comput Sci Electron (IJETCSE)
11(5):18–21
11. Pandey J, Kazmi SIA, Hayat MS, Ahmed I (2017) A study on implementation of smart library
systems using IoT. In: 2017 international conference on infocom technologies and unmanned
systems (trends and future directions) (ICTUS), Dubai, December 2017, pp 193–197. https://
doi.org/10.1109/ICTUS.2017.8286003
12. Sowah RA, Ofoli AR, Krakani SN, Fiawoo SY (2017) Hardware design and web-based
communication modules of a real-time multisensor fire detection and notification system using
fuzzy logic. IEEE Trans Ind Applicat 53(1):559–566. https://doi.org/10.1109/TIA.2016.261
3075
13. Vieira A et al (2019) Smart campus parking – parking made easy. In: Rodrigues JMF, Cardoso
PJS, Monteiro J, Lam R, Krzhizhanovskaya VV, Lees MH, Dongarra JJ, Sloot PMA (eds)
Computational science – ICCS 2019, vol 11540. Springer, Cham, pp 70–83
14. Bagula A, Castelli L, Zennaro M (2015) On the design of smart parking networks in the smart
cities: an optimal sensor placement model. Sensors 15(7):15443–15467. https://doi.org/10.
3390/s150715443
15. Popescu DE, Bungau C, Prada M, Domuta C, Bungau S, Tit DM (2016) Waste management
strategy at a public university in smart city context. J Environ Prot Ecol 17(3):1011–1020
16. Visvizi A, Lytras MD (eds) (2019) Smart cities: issues and challenges: mapping political, social
and economic risks and threats. Elsevier, Amsterdam; Oxford; Cambridge
Collaborative Metacognitive Technique
in Smart Tutoring System
Abstract The smart tutoring system is a system that helps learners to solve their exer-
cises. To improve this system, an integration of collaborative technique is required.
This tool, aid learners to understand very well their courses and resolve the difficul-
ties of case studies that are based on team work. The aim of this paper is to fix the
problem of the implicit tutor in the smart tutoring system and propose the integra-
tion of collaborative metacognitive technique as a solution. The process starts with a
collaborative design phase and finished with the implementation phase using activate
storyline 360 environment.
1 Introduction
Online courses can meet many needs such as group work, solving problem skills, etc.
The advantage of this type of training is that it can train as many learners as possible
in the entire world at their own place, to promote collaboration. A lot of researchers
suggest that the collaborative learning can improve the teaching [1]. For example,
researchers had detected that collaborative learning and the quality of learning of the
experimental group was considerably higher than that of the monitoring group [2].
Nevertheless, some learners don’t want to rely on to any group or contribute their
works [3]. Others may feel unjustly exploited by members of the group and may give
up on their efforts to them because of the sucking effect [4].
These problems are due to the fact that there is no appropriate tool to supervise and
quantify the learner’s process in collaborative learning. The researchers suggested
different contributions to promote the engagement of learners from different perspec-
tives [5]. Some researchers are concentrated on the feedback model of the smart
tutoring system using analytical mechanisms to take into consideration the level and
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 643
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_58
644 H. Fatima-Zohra et al.
previous knowledge of the learners [6]. Other researchers had focused on the engage-
ment of learners in teaching groups [7]. To augment the efficiency of the learning
process, they [8] proposed an adaptation of learning sequence using a genetic algo-
rithm to optimize a list of exercises in a training course. However, with the emergence
of these new teaching and learning technologies due to the pandemic of coronavirus;
the changes brought by new digital techniques are particularly relevant with regard
of the content and the quality. In this regard, the educational mediatization raises
many issues [9]. The objective of this paper is to describe a conceptual solution for
the computer modelling of educational knowledge, which should be according to the
current learning. The goal is to improve collaborative learning using metacognitive
techniques into the intelligent tutoring system. The paper is described as follows:
after presenting the background of the topic, the authors will define the concept
of collaborative learning and its main benefits. After that, they will describe the
proposed process of integrating Collaborative metacognition technology in the intel-
ligent tutoring system. Finally, they will present the implementation phase of the
proposed process.
2 Background
In this section, the authors will present: some techniques related to this work; the
collaborative design phase and inquiry based learning.
Students explore, observe, plan and perform different experiments, modify vari-
able values, make predictions and interpret results. Exploration is a systematic
method for conducting a search with the intention of finding the relationship between
the variables that are involved.
Interpretation of data aims to find the meaning of data collected, by synthesizing
new knowledge. The final result of the research phase is the following interpretation
of the data (discovery of the relationship between the variables) that answers the
main problem.
The Metacognition make it possible for people to take care of their own learning.
It is about increasing their sensitivity to how they learn, measuring their learning
needs, building strategies to meet those needs, and then applying those strategies.
The existing Intelligent Tutoring System contains three modules: Student module,
Expert module and Instructional module. This later contains two types of tutors
which are: Explicit and implicit tutor. Our goal in this contribution is to improve the
implicit tutor using a specific type of indirect intervention which is: Collaborative
metacognitive techniques; this method involves the conscious monitoring of cogni-
tive strategies to accomplish particular goals. The Fig. 2 present the description of
the contribution.
Collaborative Metacognitive Technique in Smart Tutoring System 647
In this section, the authors describes the process of integrating collaborative metacog-
nitive methods into smart tutoring system, and how the inquiry based learning and
conceptual map can be used to improve the interactivity in online collaborative space.
The process illustrated in the Fig. 3 present the indirect interventions used by the
implicit tutor in the instructional module of smart tutoring system. The steps that
were presented in the picture are described as below: The learners connected to an
E-learning platform. After that, they subscribe into the online collaborative space;
the learners who passed a pre-Test, a list of activities are enabled such as: resolving
the study case with their team, group activities, etc. The implicit tutor intervene
using the Collaborative Metacognitive Technique, including inquiry based learning
method and conceptual maps tool.
The Fig. 4 presents the effectiveness of meta-cognitive technique. N indicates
the total gross number of statements in each category. The line is corresponding to
a statement that meets the criteria. Collaborative meta-cognition is marked with an
asterisk [14].
648 H. Fatima-Zohra et al.
Fig. 3 The process of integrating collaborative meta-cognitive technique into smart tutoring system
based learning as described in the didactical phase. After that, the authors build a
storyboard of the proposed approach and then, they develop a solution using LMS
to create the E-Learning platform.
The Collaborative design phase is related to the proposed approach of this contri-
bution and the simulation phase. The essential phases of this collaborative design
illustrated in the Fig. 5 and described on details as below:
• Pedagogical phase: present the appropriate approach used in the contribution;
since the authors will use meta-cognitive strategies, a collaborative workspace is
required.
• Didactical phase: contains a list of tools to apply the pedagogical phase which is
the uses of the meta-cognitive techniques.
• Mediatic phase: In this phase the authors refer to the storyboard to carry out the
pedagogical activities, exercises, graphics, etc. it allows them to test the module
to identify and address any ergonomic problems.
• Technological phase: the aim of this phase is to develop interactive interfaces.
5 Implementation Phase
Articulate storyline is an authoring tool that allows the creation of interactive courses
and E-learning modules without the necessity of using programming language.
Indeed, this software allows different participant in training to easily produce inter-
active content. This tool contains four versions, but we will describe the three recent
versions that are presented as below:
650 H. Fatima-Zohra et al.
• The first one is: Articulate storyline360: which contain a list of product, some
of them will be downloaded and others will be installed, also it requires a yearly
subscription;
• The second one is: storyline3: it has different states that allow us to modify the
appearance of an object according to the learners’ actions. Storyline3 also provide
an enlarge button when the learner hovers over it or add a highlight effect, and
when he/she click on it. States also make it possible to modify the expression and
attitude of a character;
• The last one is: Storyline2: it offers an improved user interface and interactive
learning element;
All the articulate storyline versions are required Microsoft.Net Framework, Visual
C++ and Adobe Flash Player (10.3). HTML5 and JavaScript are needed also. The
articulate storyline versions contain the same options. However, there is some distinct
point that makes us easy to choose the appropriate version. The Fig. 6 represents a
comparative study for the three versions of articulate storyline to implement our
pro-posed process:
The learners will subscribed and build a team called “Our Team”. Then, they
will forward to an interactive collaborative space which they should check an option
from the list illustrated in the Fig. 8. After that, if for example the learner selects the
option: group activities; the case study will appear in Fig. 9. This activity contains 26
min, if they could not resolve, or they had a problem. The tutor will intervene with
two metacognitive techniques: inquiry based learning and mind mapping which can
help them to sort out the case study.
6 Conclusion
This paper presents the issues of classical E-Learning platform. The authors under-
took the collaborative learning: definition and their advantages. Then, they presented
the related works, after that, they described the contribution, presented the problems
of the implicit tutor in the smart tutoring system and proposed the solution which is
the integration of the collaborative metacognitive techniques.
The proposed solution can’t be realized without an essential process which is
the collaborative design phase. After the pedagogical design, a simulation of the
proposed approach is presented in this paper. The integration of a smart technique in
collaborative learning in a corporate context will be done with an efficiency result.
However, the practical implementation of these generic recommendations and the
validation of their effectiveness are deferred to future work. These experiments will
be carried out on the e-learning platform of the Larache Polydisciplinary Faculty,
which is part of the Abdelmalek Essaadi University in Morocco.
Collaborative Metacognitive Technique in Smart Tutoring System 653
References
Abstract The emergence of MOOCs (Massive Open Online Course) increased the
need for personalized student support. Question and Answering solution can reduce
the dropout rate and improve learning. In this paper, we explore the potential of using
chatbots to enhance the learning experience for students and how to take advantage
of Deep Learning. A review of chatbots models suggests that task-oriented ones
are more convenient in the educational context. To address this issue, we propose
an integrated chatbot in Open edX platform. We study architecture and proposed
technical options. The model is developed to perform a question-answering task. The
trainings were carried out using TensorFlow and NTLK. The paper also presents the
various steps in the conception and implementation. The chatbot performs well and
contributes to the improvement of the e-learning systems.
1 Introduction
Morocco has made significant efforts to generalize access to the education system.
Unfortunately, this system suffered from school dropout. According to UNICEF, the
school enrollment rate in rural areas does not exceed 32.4%. This rate drops to 30.6%
for access to high school and does not exceed 21.9% for girls [1]. The reason for
this loss is the distribution of villages over vast areas and the geographic difficulty.
UNICEF recommends the use of innovative technologies in parallel with traditional
training to fight school dropout. Indeed, the rural environment can benefit from the
new e-learning tools.
With the Sars-Cov-2 pandemic, the educational offer is becoming increasingly
restricted. Indeed, the constraints of distance oblige us to adopt new learning styles.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 655
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_59
656 R. Karra and A. Lasfar
The classroom is, by definition, a closed space and a rapid transmission vector of
Sars-Cov-2. Adaptation necessarily requires the adoption of modern technologies.
Blended-Learning is one of the effective solutions in the administration of this
pandemic. Shifting between face-to-face in small groups and remote study seems
like a good compromise. This allows one group of students to study in class and the
other at a distance. This alternation yields the possibility of taking advantage of both
methods. Since several universities have put their courses online, the flipped class-
room becomes a desirable choice. The knowledge acquired by the student during the
distance study can be deepened or corrected with the teacher. All these instructional
strategies need improved distance learning tools.
Artificial intelligence is redefining e-learning. From preventing student’s dropout
to aiding and providing information about the university and the courses, there is a
lot of ways how AI can help [2]. Universal Design for Learning (UDL) is a scientific
framework for education. It aims to improve and optimize methods in education. It
is based on how students interact with educational systems and derive from research
in education and neuroscience. It recommends offering more than one format for
learning and multiple ways to interact with content [3].
It is difficult to monitor a hundred students at the same time and offer support
to each of them. One of the options we explored is adding a chat agent that allows
students to get automatic answers.
Whereas many economic sectors have benefited from chatbots capabilities.
The education sector has lagged. Chatbots can supply opportunities to improve
engagement, proximity, and assistance between student and teacher.
Making a chatbot can be done in two ways, the rule-based approach or the base-
ruled approach. The second one makes the system learn from data. One of the
best known is the ALICE system, which is based on Artificial Intelligence Markup
Language (AIML). In recent years, several chatbots have emerged, especially those
based on a natural language process engine (NLP) and hosted in the cloud such as
Cortana from Microsoft or Alexia from Amazon [4].
NLP uses neural networks to understand what humans intend to tell (the inquiry).
Several types of chatbots exist including the question answering bot. Q&A bots are
trained on a knowledge base. However, the knowledge base must be quite substantial
to give reliable results. For NLP to perform well and give satisfactory results the
knowledge base must be large and complete [5]. To implement a chatbot successfully,
we must identify the target audience, its behavior, and the use cases. Knowing the
target audience gives us the decision on how the chatbot responds to question either
in a friendly or professional way. Ideally, the chatbot should be able to engage in a
human discussion according to the chosen topic.
A Q&A system is a combination of NLP model and database knowledge. Q&A
bots make it possible to deal with scenarios never considered and generate responses.
In these chatbots, the synergy between the system and the student is typically a one-
to-one sentence. Some students do not adapt to the transition from high school to
university. Indeed, they do not manage to follow a system of self-education [5]. The
chatbot (Q&A) finds its importance in lectures where the number of students easily
exceeds a hundred and the teacher cannot answer all the questions [6].
Enhancing Education System with a Q&A Chatbot … 657
2 Background
Deep Neural Networks (DNNs) is one of the most efficient machine learning models.
He achieves excellent results in problems as difficult as diverse like the conception
of new drugs or facial recognition [7]. Despite these advantages, DNNs only apply to
issues where inputs and outputs can be encoded to fixed sizes [7]. They accept a vector
of fixed size (question) and give a result of fixed size (answer). This limitation restricts
the use of DNNs, especially with the problems of variable sequences (dimension and
length).
Recurrent Neural Network (RNNs) is a generalization of neural networks with
sequential characteristics. They find nonlinear relationships between sequential data
of input and output. For an input sequence (I 1 , I 2 … I m ) the RNN calculates the
sequences of hidden layers (h1, h2 … hm ) and the output sequence (O1 , O2 … Om )
by iterating the function:
h t = tanh W hx It + W hh h t−1 (1)
Ot = Woh h t (2)
With tanh nonlinear function that act as a coordinate-wise and W weight matrix. In
practice, RNNs are confronted to the instability problem due to gradient descent. This
causes the problem of vanishing or exploding gradients. The difficulty of training
RNNs for problems with long-range temporal dependencies almost hampered their
development. As shown in (Fig. 1), Sequence-to-Sequence model (Seq2Seq) is based
W X Y Z <eos>
Encoder
<eos> W X Y Z
A B C Decoder
The Q&A chatbot simulates a conversation between the student and a virtual
instructor. Chatbots operate on a one-to-one logical sequence [10]. The informa-
tion flow is a question followed by an answer. A student enters a question and waits
for the chatbot to give a correct answer.
Open edX project is a web application based on the Django framework for
producing MOOCs. Django powers edx.org and many other leading online education
sites. It is based on two main components, the studio, and the learning management
system. The studio can be considered as a CMS (Content Management System) for
course authoring, scheduling, and grading policy [11]. The LMS is the foreground
of the Open edX project. Learners use it to take courses and exams. The Open
edX system is based on a modular architecture called XBlocks. It makes it possible
to develop new functionalities for educators and learners. The XBlock technology
allows the use of several components from the Open edX platform. The different
applications are grouped so that they can interact with each other and with the main
Enhancing Education System with a Q&A Chatbot … 659
system. XBlocks can be used to represent a single module, web component, multi-
media elements, social learning experiences, or interactive content. Furthermore,
sections or entire courses can be constituted with the XBlocks.
Open edX uses many technologies. A basic XBlock has fields, a student View,
handlers for web services, and the web pages where the results are rendered.
MongoDB is used for course persistence and data analytics [11]. MySQL database
to store user data, settings, and their progress. (Fig. 2) shows a general description
of the Q&A chatbot:
Our Chatbot extension has been included as a new plug-in inside the LMS.
XBlock’s front-end use Bootstrap. It makes our Q&A compatible with mobile
devices. As part of the entire system, we have implemented a Q&A XBlock equipped
with an NLP model capable to retrieve answers and interact with students. This model
uses different tools principally TensorFlow 1.13.1 and NLTK 3.5 for the DNLP capa-
bility. The Q&A solution was trained with GPU support using Nvidia Quadro P3000
6GB. There are different training strategies. [5] trained the chatbot on a first dataset
to give it the language capabilities and on another dataset about insurance so it has
domain-specific knowledge. Our approach is to separate trained models. Each course
Learners
ChatBot
Deep Natural
Processing and rules parser
XBlock
Language Processing
TensorFlow
Tokenization
Numpy
NLTK
Seq2Seq
Text processing
Encoder Decoder
Entity extraction
RNN-LSTM
Intent classifier
Knowledge Base
Data Source
MongoDB
has its own model [12]. With this choice, we avoid the catastrophic forgetting effect.
The loss of previously gotten knowledge when we try to train our model on more
data. In fact, [4] have shown that the forgetting phenomenon is not reserved for DNN
but also LSTM. Their model was able to perform an accurate output only on the
newly trained sequences. The base knowledge is persisted in MongoDB (already
provided by Open edX platform) as collections of BSON documents. (Fig. 3) shows
the proposed data model for discussions:
Course resources must be formatted following the ‘training_resource’ collection
of four fields _id, course_id, question, answer. The history of discussions between the
bot and the student are stored in the collections question_answering(_id, course_id,
user_id, discussions) and discussions(_id, text_disc, type_disc, date_disc).
To develop the chatbot we used the XBlock-SDK. It is a set of features and rules
to follow. The developed subsystem is compatible with Open edX. The Xblock-SDK
has the same runtime as the whole Open edX solution [10]. These specifications make
it very convenient for development and test purposes. In a production environment,
we can change the XBlock default database from SQLite to MongoDB.
4 Results
Integrating the chatbot with the LMS appears to be a promising solution for improving
the quality and interactivity of the e-learning system. During the tests, we have
asked several questions to our ChatXBlock. The responses were generally satisfying.
The user interface is designed in an intuitive way. It also respects the Open edX
graphic design. Simplicity was one of the pillars of our design. The student gets the
information intuitively.
(Figure 4) shows the process of dialog between the chatbot and a student enrolled
in a course:
However, we found some limitations in the Q&A chatbot. If the student asks badly
or makes a mistake in the question, the chatbot gives an incoherent answer. The
second limitation is how to guide the student to ask the right questions. Especially
if we assume the student course ignorance. The chatbot is useful only in the case of
oversights or revisions. Indeed, it does not help in the first phases of learning where
the student begins to discover the subject, and even worse can be misleading.
Enhancing Education System with a Q&A Chatbot … 661
Human Takeover was added to the system. It means the taking over of the conver-
sation by the instructor if the chatbot does not respond. This possibility is available
when the response of the chatbot does not satisfy the student. He can trigger the
intervention of the instructor [6]. This allows the student to come back to the teacher
if necessary and verify the chatbot answer. The mixture of machine and human
interaction enhances collaborative teaching and gives diversity in learning styles.
5 Conclusion
Several disciplines overlap for the enhancement of the student experience. Students
need feedback and guidance. In this study, we presented the challenges of recent
technologies for education. We have developed an excellent Q&A chatbot by using
deep learning and natural language processing. It fills the lack of proximity and
availability of the instructor in MOOCs.
It remains appropriate to focus our studies on knowledge transfer. It allows adding
blocks of knowledge to a pre-trained model without going through new training each
time. This way is closer to the natural functioning of the human brain and reduces
the duration and cost of each new training. [13] has introduced two parameters to
662 R. Karra and A. Lasfar
control the user’s recommendations according to their preferences and find a balance
between questions on one side and recommendations and their relevance on the other
side. Extending our chatbot with recommendations possibility is another issue that
worth considering. The chatbot can recommend a module or a chapter that matches
a user’s gap based on the questions the student posed.
References
Abstract E-learning has recently become widespread all over the world. All coun-
tries are working more and more to integrate information and communication tech-
nology in their educational systems. In this concern, recommender systems have
played a crucial role for suggesting items that best match user’s preferences in
different area. Currently, in order to prevent the spread of coronavirus pandemic and
to ensure continuity of education during the confinement period likewise our country
Morocco, distance learning has become an imperative necessity. In this regard, we
are going to delve into the role of e learning in achieving quality education and the
difficulties of its implementation in Moroccan context. Several digital courses have
been published on e-learning platforms. However, learners, who do not have enough
experience to choose the courses they need, often face many challenges. In this paper,
we present Multi-agent system as a helpful tool in recommendation, then we review
the current recommendation methods. Moreover, we implement a recommendation
system for e-learning platforms that recommends university courses for students.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 663
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_60
664 M. Amane et al.
at the same time, they can manage this crisis successfully without blocking the
educational system. Nevertheless, it has so many challenges; particularly, in the time
consuming for searching a good support. Recommendation systems can be one of the
best solution to this problem for learners, it can help them to find their favorite content
from the huge amount of information [1]. Most of the choices that individuals select
are constructed on recommendations or suggestions from others [2]. For example, in
e-commerce, clients need to examine item surveys before buying. Netflix.com uses
the recommendation to assist individuals discover movies and videos that interest
them using collaborative filtering; Youtube.com creates a personalized home page
which displays the videos recommended for each user according to his access history
[3]. In the same way, in the education field, a recommendation system is a great tool
that can help to improve learning. For instance, in university, RS could help students
to find suitable courses quickly as well as to reduce the time to explore the courses
they will take [3]. The learner can have an overview of the course according to the
vote, trends and content of the course, so the analysis of the best course can be carried
out easily. Many online learning platforms have been designed to improve access to
high quality educational content for students connected to Internet [4]. To provide
suggestions to learners who do not have enough experience to choose the courses they
need, researchers have tried to develop various approaches to recommend relevant
learning objects to learners in order to improve the effectiveness of studies. In fact, it
has become ultimately a pre-requisite to implement our recommendation e-learning
system in intelligent way, thanks to the significant role that recommender systems
play in improving teaching and learning quality. In this concern, and in order to
fulfil communication and interaction between users, we are going to use multi-agent
systems to improve the quality of the courses recommendation list, by extracting
positives and negatives comments of each course, and then reclassify the final list.
This method can enhance and recommend the best courses to learners.
This paper is oriented to the use of recommendation systems in online learning
systems, it is organized as follows: In sect. 2, we provide details about the common
used recommendation methods. In Sect. 3, we explain the proposed approaches
by researchers in e-learning recommender system, as well as the recommendation
techniques used in each article. In Sect. 4, we present our proposed course recom-
mender system, using the content-based approach and the multi-agents system, the
description of course and the similarity.
system. At the point when you select a product to purchase, Amazon suggests other
products that different clients bought dependent on that original product [6]. In the
context of e-learning, in a course recommendation system for example, the system
can suggest a course to a learner without needing to know the content of the course,
but only based on the ratings of similar learners.
3 Related Works
4 Discussion
Most of the approaches proposed above use collaborative filtering as part of their
recommendation systems. This technique can be used to create recommendation
systems that give suggestions to a student based on the interests and preferences of
similar learners, one of the advantages of collaborative filtering is that they do not
need or do not require any knowledge of database elements, the user ratings matrix
is the only entry required. However, it has various drawbacks, such as:
• Cold start problem: It occurs when the recommendation system is unable to
make decisions about users or items, on which enough information has not been
collected. The first time a new user visits a specific recommendation system,
for example, none of the items have been rated. Therefore, the recommendation
system does not know what are the likes and dislikes of this user. The same problem
occurs when a new item is introduced to the recommendation framework. Since
nobody has ever evaluated this element, the recommendation system cannot know
A Multi-agent and Content-Based Course Recommender System … 667
what other elements that are similar. For example, in a course recommendation
system, new courses cannot be recommended until certain number of learners
evaluate it [23].
• Sparsity: is also a significant problem in collaborative filtering, it occurs when the
number of learners who have rated courses is too small compared to the number
of available courses. In practice, it is difficult to find similar neighbors for the
learners to generate appropriate recommendation. This is one of the most obvious
drawbacks of recommendation systems based on collaborative filtering [23].
To deal with these challenges, we propose to use content-based approach with
multi-agent system for our course recommender system, because this method can
overcome the cold start problem, thus, it only has to analyze the items and a single
student’s profile for the recommendation.
668 M. Amane et al.
5 Methodology
Two primary stages of calculation are involved in the vector space model: weight
calculation and cosine similarity calculation [11].
• One of the most used CBF techniques is the TF-IDF measure [24]. The weight
wi,j is calculated (TF-IDF) with normalized frequency as follow in Eq. 1:
f i, j D
Wi, j = × log (1)
max t f t, j di
where:
– fi,j: number of term i appear in text j,
– maxt ft,j : is the Max frequency by the whole of all t keywords that occur in d j,
– D is the sum of recommended documents to learners, and di is the number of
documents that comprise term i.
• After the weighting step, a measure of vector similarity must be defined. This
similarity aims to determine the proximity between two documents. There are
many similarity measures, but the most widely used measure in the literature is
cosine similarity [11]:
Where:
→ − −
→− →
Cos − → = wc . ws
wc , w (2)
wc .−
−
→ →
s
w s
670 M. Amane et al.
To test the recommendation systems, the datasets are available on MovieLens [11],
these data are less suitable for testing on the recommendation systems in e-learning
because the datasets are based on the film rankings. For the experiment, we use
a sample of dataset of university courses, which contain information (Course Id,
Course Title, Description…) about 8012 courses as shown in this capture (Fig. 4).
After preprocessing the data and choosing the relevant features for our problem,
which are Course Title and Description, we start calculating TF-IDF scores, and
we compute the Cosine Similarity between courses. The next step is to returns the
most similar courses based on the cosine similarity score. After developing and
implementing our recommendation system, we have as result the list of recommended
courses.
For example, the top ten similar courses of “ASP.NET 3.5 Fundamentals” are
(Table 2):
References
13. Ray S, Sharma A (2011) A collaborative filtering based approach for recommending elective
courses. In: Dua S, Sahni S, Goyal DP (eds) Information intelligence, systems, technology and
management. Springer, Heidelberg, pp 330–339
14. Taha K (2012) Automatic academic advisor. In: Proceedings of the 8th IEEE international
conference on collaborative computing: networking, applications and worksharing. IEEE, Pitts-
burgh
15. Nagpal D, Kaur S, Gujral S, Singh A (2015) FR: a recommender for finding faculty based on
CF technique. Procedia Comput Sci 70:499–507
16. Shu J, Shen X, Liu H, Yi B, Zhang Z (2018) A content-based recommendation algorithm for
learning resources. Multimed Syst 24:163–173
17. Bhumichitr K, Channarukul S, Saejiem N, Jiamthapthaksin R, Nongpong K (2017) Recom-
mender systems for university elective course recommendation. In: 2017 14th international
joint conference on computer science and software engineering (JCSSE). IEEE, Nakhon Si
Thammarat, pp 1–5
18. Turnip R, Nurjanah D, Kusumo DS (2017) Hybrid recommender system for learning mate-
rial using content-based filtering and collaborative filtering with good learners’ rating. In: 2017
IEEE conference on e-learning, e-management and e-Services (IC3e). IEEE, Miri, Sarawak, pp
61–66
19. Khoja Z, Shetty S (2017) Hybrid recommender system for college courses. In: 2017 interna-
tional conference on computational science and computational intelligence (CSCI). IEEE, Las
Vegas, pp 1167–1171
20. Wang G, He X, Ishuga CI (2018) HAR-SI: a novel hybrid article recommendation approach
integrating with social information in scientific social network. Knowl-Based Syst 148:85–99
21. Bhagirathi YN, Kiran P (2019) Book recommendation system using KNN algorithm. 2:4
22. Sunil, Doja MN (2020) An improved recommender system for e-learning environments to
enhance learning capabilities of learners. In: Singh PK, Panigrahi BK, Suryadevara NK, Sharma
SK, Singh AP (eds.) Proceedings of ICETIT 2019. Springer International Publishing, Cham, pp
604–612
23. Jain S, Grover A, Thakur PS, Choudhary SK (2015) Trends, problems and solutions of recom-
mender system. In: International conference on computing, communication & automation.
IEEE, Greater Noida, pp 955–958
24. Sielis GA, Tzanavari A, Papadopoulos GA (2015) Recommender systems review: types, tech-
niques and applications. 12
Image and Information Processing
Sentiment Classification for Moroccan
Arabic Using the Lexicon-Based
Approach
Abstract Social networks users contribute in producing huge quantity of data due
to their increasing online activity. However, extracting useful knowledge from social
networks can be a difficult process, especially when conducting sentiment analysis
on Arabic texts. The challenging aspect of Arabic language resides primarily in the
mutual use of standard and dialectal forms in expressing opinions. This work focuses
on sentiment analysis using the semantic orientation approach that rely on a lexicon
of opinion words and a rule-based classifier to achieve sentiment classification. In
this methodology, we build a large sentiment analysis lexicon that covers modern
standard Arabic and Moroccan dialectal Arabic. Then, we have developed several
rules and tested them in the classification process. The proposed methodology showed
promising results.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 675
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_61
676 C. Ferroud and A. Elouardighi
In sentiment analysis (SA), we intend to find out attitudes, knowledge and feelings
declared in a shared text or comment, in order to extract its polarity: positive, negative
or neutral. Two approaches have been developed for SA. One based on the machine
learning (ML) and the other one on semantic orientation (SO) [3].
In this work, we focus on the SO approach, an unsupervised sentiment classifica-
tion based on a lexicon and rule-based algorithm, in order to classify Arabic opinions
that are expressed in modern standard Arabic (MSA) and Moroccan dialectal Arabic
(MDA). Thus in the first step of our methodology we build a lexicon that covers
both languages, and then we develop a set of rules that enable to calculate the SO.
Moreover, we want to test the effect of different polarity rating scales on the SO.
Firstly, the binary rating scale; where each word in the lexicon is rated with +1 if it
is positive or −1 if it is negative. Secondly, the ordinal rating scale; where each word
is rated in a scale from +1 to +5 if it is positive and from −1 to −5 if it is negative.
We start our article by presenting the most relevant work of the SA employing
the social networks, then we describe our SA approach mainly the construction of
a SA lexicon in MSA and MDA as well as the design and implementation the rule-
based classifier. The main obtained results are presented in Sect. 4, and we close our
contribution with a conclusion and some possible perspectives for this work.
2 Related Work
Over the past decade, there has been an explosion of research works interested in
SA such as analyzing attitudes and emotions of individuals towards entities such as
products, services, organizations etc. We start our paper by presenting some research
work in this area and close to our contribution.
In [1] Abdulla et al. proposes two approaches for SA: one based on the Arabic
corpus and the other on the lexicon. As there are few publicly available Arabic
datasets and lexicons for SA, they created by hand an annotated dataset and they
presented the steps of building their lexicon. They found that for their first approach,
the SVM method used for classification offers the highest precision.
In addition, they noted that the increase of the lexicon allows an improvement in
the performance of their second approach.
Hatzivassiloglou and McKeown [2] used a method that recovers automatically
the SO information positive or negative by collecting indirect information from a
large corpus of the conjoined adjectives. To predict whether they have the same or
different SO, they used a log-linear regression model. The achieved quality of the
developed model was 82%. They classified several sets of adjectives based on the
inferred links and labelled them as positive or negative. They achieved 92% accuracy
for the classification task and 100% for the labelling one.
Duwairi [4] used ML techniques for SA to define the polarity of Arabic tweets
with dialects. One of his dares is that stemming algorithms do not work well with
dialects words with an extended set of stop words. He proposed a Framework that
is able to perform SA on tweets written in MSA or in Jordanian dialectal Arabic.
Sentiment Classification for Moroccan Arabic … 677
The basis of this Framework is a dialect lexicon that maps dialect words to their
corresponding MSA words.
Hu et al. [5] used a SA approach that depends on the correlation and indication
of emotional signals and was tested on two sets of twitter data. The obtained results
show the effectiveness of their approach and the significant of including the different
emoticons in the SA.
From a large data set collected from Arabic tweets, Twairesh et al. [6] have
developed two lexicons for SA. The importance of their lexicons is that their capacity
to capture the idiosyncratic nature of social media text. In addition, the possibility
of using them in different kinds of issues such as product reviews.
Ibrahim et al. [7] used a characteristic-based sentence-level approach for Arabic
SA. The approach developed is based, on the one hand on the lexicon of Arabic
idioms/sayings to improve the detection of the sentiment polarity in Arabic sentences.
On the other hand, syntactic features were used for the conflicting sentences allowing
the improvement of the performance of the sentiment classification. They introduced
a wide lexicon of Arabic sentiment words and automatically expandable. The lexicon
building process started by a manual step that consisted of collecting and annotating
gold-standard sentiment words, then it was expanded automatically through word
research in an online Arabic dictionary. The data set focuses on tweets and microblogs
written in modern standard Arabic and in Egyptian dialectal. The experimental results
indicated high levels of performance, with a preciseness over than 95%.
Abdul-Mageed and Diab [8] presents SANA, a huge Arabic lexicon that is a large-
scale multi-gender lexicon for Arabic SSA (subjectivity and sentiment analysis). This
lexicon covering also, the Egyptian Dialect and the Levantine dialect Arabic, and
was developed using manual and semi-automatic techniques.
Al-Ayyoub and Bani Essa [9] developed a sentiment lexicon with nearly 120,000
Arabic terms and a tool based on calculating predicates for AS using the lexicon
approach, which was one of the most commonly used technique. The obtained results
were better in terms of accuracy in comparison with that based on keywords approach.
Khaliq and Carroll [10] used an unsupervised methodology to learn non- concate-
native morphology, which they apply to induce a standard Arabic lexicon of roots
and pattern models. Their approach was based on the concept that roots and patterns
can be affirmed by jointly recursive notation based on hypothesized patterns and root
frequencies. Their morphological analysis with the induced lexicon carry out a root
identification precision greater than 94%.
Shoukry and Rafea [11] developed a sentiment classification model for the Egyp-
tian dialect using Arabic tweets that are analyzed to identify their polarity (positive or
negative). Their model finds its importance in the region of the Middle East that uses
largely the Arabic in the twitter network. Overall, this SA works has dealt with Arabic
language based on two approaches, the ML approach that uses supervised classifiers
and the SO approach that requires the construction of a lexicon. In the SO approach,
the authors followed different methods to construct their lexicons, including manual
and automatic methods. In addition, some works have been limited to treating the
standard Arabic only, while others have extended their lexicons by integrating one
or more Arabic dialects. In our work, we focus on SA to deal with comments written
678 C. Ferroud and A. Elouardighi
in MSA and in MDA, indeed this dialect is still not fully addressed in the context
of the SO approach. This is why the construction of a new lexicon devoted to both
standard Arabic and the Moroccan dialect constitutes an essential element of our
contribution.
Supervised leaning models built for SA require a labelled data for learning. However,
the manual sentiment annotation can be very costly in time and effort. In addition,
the constructed model may be biased by the domain [12]. We devote this section
to present our methodology for polarity classification that address the sentiments
expressed in standard Arabic and Moroccan dialect.
The first step towards building our Arabic lexicon for SA is the collection of seed
words, which have been the subject of several extensions using synonyms of words
taken from an online dictionary [16]. For each word, about two to three synonyms
Sentiment Classification for Moroccan Arabic … 679
was added by inheriting the same polarity as their original seed words, and we have
translated the standard Arabic lexicon into the Moroccan dialectal.
In the end, the lexicon contained 5942 words, with 2127 positives word and 3815
negatives. To obtain the SO of an Arabic comment, we have built a tool that associate
each word of the lexicon to a weight identifying its polarity. The word does not exist
in the lexicon its polarity is null. For feature construction, we have used the n-gram
method that each word is considered as one token regardless to the surrounding
words. To calculate the SO of the text, it suffices to aggregate the polarity score of
each word constituting the text from the lexicon.
The obtained score represents the polarity of the whole text. In order to enhance
the performance of our lexicon, we have implemented some rules for negation,
intensification and a third rule for conflicting phrases.
A negative value is assigned to words that inverse the polarity of the word [15, 17].
The following list presents some negation words in MSA and MDA:
For example in (I don’t like noise) the verb (I like) is positive
word in our lexicon with polarity equals to 1. However, due to the presence of the
negation word( الdon’t) which precedes it, switch the polarity to −1. In another
situation, negation word reverse completely the polarity of any word to the opposite.
In addition, we have enriched the list by inserting some negation words from the
Moroccan dialects such as .
On the other hand, intensification increases the intensity of words polarity to make
it even stronger [17, 18]. Some words were used for intensification:
. Our approach deals with such cases by adding +1 to a positive word and
−1 to a negative one. For instance, the sentence (This book is very
cool), has a polarity equals to 2 since has +1 in our lexicon, and the presence of
(very) accentuate the polarity by append another +1.
Conflicting phrases consist of two words with opposite polarity; noun followed
by adjective, one of them is positive and the other is negative [7]. For example, the
pattern (Mysterious vision) consists of two sentiment words, the noun
(vision) which express positive polarity and the adjective (Mysterious)
which express negative polarity. Using the lexicon only, their polarity is 0, but our
classifier labelled it as negative polarity. As we observed in most cases, if there exists
a pattern of two opposite sentiment words, it yields a negative sentiment polarity
such as (Temporary release), (fake love), (misbehavior)
680 C. Ferroud and A. Elouardighi
Fig. 1 Sentiment classification process of Arabic comments using the lexicon-based approach
Table 2 Example of polarity score of Arabic comments with binary and ordinal scales
4 Experimental Results
We dedicate this section to present the main results of the experiments obtained as
well as their comparisons with those obtained in other similar works. Two dataset are
used for testing the efficiency of our methodology. The first one is a balanced dataset
that contain 5000 labelled comments (2500 positive comments and 2500 negative
ones) and concern various domains including politics, culture, sports etc. The second
dataset is unbalanced since it contains 3544 positive comments and 6400 negative
ones with a total of 9944 comment [19]. Both datasets contain comments written in
MSA and MDA.
Sentiment Classification for Moroccan Arabic … 681
Table 3 Classification
Dataset: Facebook 5000
results: dataset 1
Steps Accuracy F-score
1- Binary scale lexicon 69,48% 69,47%
2- Ordinal scale lexicon 72,41% 72,41%
3- Ordinal scale lexicon + rules 73,05% 73,04%
Table 4 Classification
Dataset: elec2016
results: dataset 2
Steps Accuracy F-score
1- Binary scale lexicon 67,04% 62,34%
2- Ordinal scale lexicon 68,31% 65,41%
3- Ordinal scale lexicon + rules 69,09% 66,04%
682 C. Ferroud and A. Elouardighi
5 Conclusion
This paper addressed Arabic Sentiment Analysis through the unsupervised approach.
The contribution consisted of building a large Arabic lexicon that contain opinion
words in Standard Arabic and Moroccan dialect. Several rules were developed in
order to achieve polarity classification of social web data. Thus, several experiments
were conducted through the various steps of this process in order to observe the
amelioration obtained on the performances. The results showed that the extended
rating scale of the lexicon (1–5) alongside with the proposed rules have returned the
best scores.
We devote the future work to deal with sarcastic comments, since detecting
sarcasm is still one of the significant issues faced in the context of SA.
References
1. Abdulla NA, Ahmed NA, Shehab MA, Al-Ayyoub M (2013) Arabic sentiment analysis:
Lexicon-based and corpus-based. In: 2013 IEEE Jordan conference on applied electrical
engineering and computing technologies (AEECT), pp 1–6. IEEE
2. Hatzivassiloglou V, McKeown K (1997) Predicting the semantic orientation of adjectives. In:
35th annual meeting of the association for computational linguistics and 8th conference of the
European chapter of the association for computational linguistics, pp 174–181
3. Liu B (2012) Sentiment analysis and opinion mining. Synth Lect Hum Lang Technol 5(1):1–167
4. Duwairi RM (2015) Sentiment analysis for dialectical arabic. In: 2015 6th international
conference on information and communication systems (ICICS), pp 166–170. IEEE
5. Hu X, Tang J, Gao H, Liu H (2013) Unsupervised sentiment analysis with emotional signals.
In: Proceedings of the 22nd international conference on World Wide Web, pp 607–618
6. Al-Twairesh N, Al-Khalifa H, Al-Salman A (2016) Arasenti: large-scale twitter- specific
arabic sentiment lexicons. In: Proceedings of the 54th annual meeting of the association for
computational linguistics (Volume 1: Long Papers), pp 697–705
7. Ibrahim HS, Abdou SM, Gheith M (2015) Sentiment analysis for modern standard arabic and
colloquial. arXiv preprint arXiv:1505.03105
8. Abdul-Mageed M, Diab MT (2014) Sana: a large scale multi-genre, multi-dialect Lexicon for
Arabic subjectivity and sentiment analysis. In: LREC, pp 1162–1169
9. Al-Ayyoub M, Essa SB, Alsmadi I (2015) Lexicon-based sentiment analysis of arabic tweets.
Int J Soc Netw Min 2(2):101–114
10. Khaliq B, Carroll J (2013) Induction of root and pattern lexicon for unsupervised morphological
analysis of arabic
11. Shoukry A, Rafea A (2012) Sentence-level arabic sentiment analysis. In: 2012 international
conference on collaboration technologies and systems (CTS), pp 546–550. IEEE
12. Read J, Carroll J (2009) Weakly supervised techniques for domain-independent sentiment
classification. In: Proceedings of the 1st international CIKM workshop on topic-sentiment
analysis for mass opinion, pp 45–52
13. Maghfour M, Elouardighi A (2018) Standard and dialectal arabic text classification for senti-
ment analysis. In: International conference on model and data engineering, pp 282–291.
Springer
14. Larkey LS, Ballesteros L, Connell ME (2007) Light stemming for arabic information retrieval.
In: Arabic computational morphology, pp 221–243. Springer
15. Saif H, Fernandez M, He Y, Alani H (2013) Evaluation datasets for Twitter sentiment analysis:
a survey and a new dataset, the STS-gold
Sentiment Classification for Moroccan Arabic … 683
Abstract The previous studies have used search query data to predict the numbers
of tourists expected to visit a particular destination. The SQD-PCA-SVR model
was established in the aim of achieving the accurate forecasts of tourism demand.
This study used Search Query Data (SQD) to supplement traditional data in order
to predict tourist arrivals in Marrakech. The Principal Component Analysis (PCA)
is applied to extract the low-dimensional and efficient feature information. Support
Vector Regression (SVR) can approximate a non-linear system. In our approach,
we used SVR with GridSearchCv and ten-fold Cross-Validation to choose optimal
hyper parameters. The SQD-PCA-SVR model is a new combination, consisting of
a search engine data, dimensional reduction algorithm, and machine learning. The
results indicate that models with SQD-PCA-SVR provide more accurate forecasts
than other models.
1 Introduction
Search query data has become an important data source for the prediction of tourism
demand [1]. Tourists are accustomed to use search engines to make different decisions
related to all details of their travels, such as hotels, restaurants and transportations.
The tourists’ queries reflect various kinds of information needs, these information
could be captured online in different time periods on those search engines.
The use of search query data in predicting tourism demand necessitates the
careful selection of keywords that have considerable correlations with the prediction
purposes. Also, high dimensions of variables may induce multicollinearity problems
[2]. To overcome the problem of higher dimension, Principal Component Anal-
ysis (PCA) is employed prior to the application of any machine learning algorithm.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 685
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_62
686 H. Laaroussi et al.
2 Literature Review
The search query data is largely used in studies to predict the tourist arrivals. For
instance, [5] used the Baidu Index and Google Trends data to forecast tourist numbers
in China by using autoregressive moving average models. [6] made use of Google
search index to predict Caribbean Islands’ tourist flows, and found that Google Trends
does in fact help to forecast tourism demand. [7] used Internet search indexes and
machine learning to forecast tourist arrivals expected in popular destinations in China.
To extract principal components from Search Query Data, PCA can be used, as it
can eliminate the problems of multi-collinearity and over-fitting. For example, [8],
by using PCA, managed to decrease the dimension of the 15 Baidu indexes that are
employed to forecast tourist arrivals of Beijing and succeeded in obtaining the first six
principal components. [2] used principal component analysis (PCA) and generalized
dynamic factor model (GDFM) to create a composite search index based on Google
trends data.
Tourism demand forecasting predominantly uses SVR, which is suitable for
nonlinear prediction. For examples, [9] combined SVR with the Bat algorithm (BA)
with the aim of forecasting tourist flows by incorporating search engine data.
Forecasting Tourism Demand in Marrakech … 687
The basic idea of SVR is as follows: Suppose training data pairs (xi , yi ) are given
where xi ∈ Rn , yi ∈ R, i = 1; 2; 3; . . . xi is the input value and y is the target value.
The SVR function can be inferred from following formula:
Where ω is the n-dimensional weight factor, b is the offset, and ϕ(x) is the map
function of mapping xi into the high l-dimensional feature space.
The final support vector regression model can be constructed using Lagrangian
function:
n
f (x) = αi − αi∗ K xi , x j + b. (2)
j=1
Where K xi , x j is the kernel function and αi , αi∗ is the Lagrangian multiplier
PCA is a frequently used data dimensionality reduction method, The PCA is used
to generate the principal components through recombination of tourism demand
influential factors. Basic steps of PCA algorithm as shown in Fig. 1.
Covariance matrix
3 Methodology
This work puts forward a forecasting framework to construct the tourism demand
prediction model using search query data, as shown in Fig. 2. The detailed steps are
as follows:
(1) Selection of the search engine platforms: In this empirical study, the majority of
travelers will opt for a search engine to obtain information, and later make the
convenient plans concerning a particular destination. In this paper, we decided
to use Google search engine data.
(2) Selection of keywords: First, we determine the seed search keywords related
to the tourist destination. Then, we procure the search volume data of each
keyword from Google Trend, and we extract the related keywords to portray
all the facets of tourists’ concerns on the destination, by means of Google
Trends related queries.
(3) Process Data: Remove irrelevant keywords using Pearson coefficient correla-
tion.
(4) Feature extraction using PCA: PCA seek to represent the variation contained
within the dataset using a small number of factors.
(5) Forecasting model: SQD-PCA-SVR model is constructed and evaluated. In
this study, a Linear Regression Model was established, as a benchmark model.
(6) Model evaluation: We used three criteria to compare the forecasting perfor-
mance of the models: Mean Absolute Error (MAE), Root Mean Square Error
(RMSE), and Mean Absolute Percentage Error (MAPE). The best forecasting
model is the one that records the lowest values on the three criteria.
SVR
SVR
5-Forecasng model Forecasng Training
6-Model evaluaon
4 Empirical Study
below this threshold are excluded. Eventually, 17 keywords were chosen. The
Pearson correlation between these 17 keywords and tourism demand is shown
in Table 2.
In this paper, the PCA method is used in the Scikit-learn machine learning library
in the Python programming language. The module sklearn decomposition provides
the PCA object, which can convert the data into principal components.
PCA is used, and 17 factors are reduced to six principal components with a
cumulative contribution rate of over 92%. The contribution rate and cumulative
contribution rate of six principal components are presented in Table 3. The cumulative
contribution rate is presented in Fig. 4.
The data procured by PCA was divided into two sections: 80% a training set and
20% a testing set. SVR model was implemented using Python and pipeline library
of Sklearn. Parameter tuning was executed by the inner-function GridSearchCV()
and ten-fold cross validation method. Grid search was applied to find the hyper-
parameter which has the best score. The best combination of parameter values is set
as follows: the penalty parameter is set to 1; the parameter gamma reaches 0.001; the
kernel function selects the radial basis function kernel; the parameter of degree is at
7; and the parameter epsilon is at 0.01. A comparison between the actual number of
tourists and the forecasted number of tourists using the SQD-PCA-SVR is illustrated
in Fig. 5. The predicted graph followed the ground truth in broad outlines.
The SQD-PCA-SVR model is compared with linear regression, for evaluation
of the two models, MAE (mean absolute error), RMSE (root mean square error)
692 H. Laaroussi et al.
Table 4 Forecasting
MAE RMSE MAPE
evaluation
SQD-PCA-SVR model 1005.35 1323.54 24.43
Linear regression model 1524.68 1976.87 30.76
and MAPE (average absolute percentage error) and were calculated using sklearn
metrics. The results are reported in Table 4. The best forecasting model is the one that
has the lowest values in these measures. Finally, the capability and efficiency of the
SQD-ACP-SVR model to estimate the tourism demand in Marrakech are validated
by the results.
5 Conclusions
References
Abstract The quantity of data produced by social networks, achieves every year
unprecedented records, because these platforms became the main sources where
subjective opinions are expressed. Processing opinions data from the raw state to
extract useful information can be a very difficult task, especially when it comes to
Arabic text that includes standard and dialectal forms. In this paper, we focus on
sentiment analysis based on Facebook comments shared in Modern Standard Arabic
and in Moroccan Dialectal Arabic. Thus, we put the text preprocessing under the
scope by testing several schemes and by inspecting their effects on the supervised
classification of sentiments.
1 Introduction
The rise and growth of social networks (SN) and microblogging has revolutionized
the social relationships and behaviors. With platforms like Facebook and twitter,
users can express their sentiments and opinions about different topics. The increasing
amount of contents generated by such platforms provide an enormous resource of
textual data and present an excellent opportunity to comprehend the user’s opinion.
Sentiment analysis (SA) is a domain of study that aim to analyze opinions and senti-
ments shared with subjective text in web platforms, such as blogs, forums and espe-
cially social networks, in order to extract valuable knowledge that can be exploited
in decision making process [1].
Because SA is based on text in natural language, the linguistic aspects reveals
the main challenges of this analysis. Particularly, the Arabic language possess a rich
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 695
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_63
696 M. Maghfour and A. Elouardighi
2 Related Work
Several studies have addressed sentiment analysis based on Arabic text; nevertheless,
the linguistic aspects still one of the major issues in SA. Thus, researchers were
intrigued by idea of investigating the effects of text preprocessing, so they have tried
to test, select and improve the preprocessing tasks in order to overcome these issues.
In their work, Shoukry and Rafea [2] carried out a sentiment analysis study based
on tweets that were published in modern standard Arabic and in Egyptian dialect;
the substantial difficulties that they have faced were about the morphology of the
Toward Improving Arabic Text Preprocessing … 697
studied dialect. For this reason, the authors have developed their own stemming
method, which is suitable for the preprocessing of texts in the Egyptian dialect.
Duwairi and El-Orfali [3] tested several preprocessing strategies for SA on two
Arabic datasets; the first concerns politics, while the second is composed from movie
reviews. Their strategies included stemming, light stemming, stop words removal,
and feature reduction. The results showed major improvement when they have applied
root stemming and feature reduction technique with every classifier: SVM, NB or
KNN.
Moreover, in [4], Duwari et al. experimented the contribution of their adopted
processing tasks in SA, namely negation detection, stemming, stop words filtering
and dialect conversion to MSA. At first, their results showed insignificant improve-
ment with stemming and stop words filtering for every classifier: SVM, Naive Bayes
or KNN. In addition, despite the expected enhancement of classification performance
following the use of negation detection and dialectal conversion, their results were
counter-intuitive since they obtained similar performance to the baseline.
Ghelli et al. [5] presented an approach to construct an annotated corpus automat-
ically for Algerian SA in Arabic and Arabizi. A corpus composed from Facebook
comments was collected. After an Algerian SA lexicon was compiled, they have
applied several ML algorithms in the sentiment classification step, such as random
forest and SVM. The preprocessing was focused on cleaning of the messages, agglu-
tination treatment and negation treatment. The obtained results showed better results
with Arabic than Arabizi and with internal dataset than the external one. This work is
more focused on testing the combined effect of feature selection with ML classifiers
and text representation schemes. The results showed the outperformance of SVM
compared to Decisions trees and Radom forest by using the same preprocessing
tasks with every classifier.
In the previous related works the sources of the collected texts are twitter, movie
reviews websites and Facebook. The use of Arabic language on these platforms is
usually associated with one dialect at least; therefore, text-preprocessing tasks are
expected to have an impact on the SA performance.
context, the presence of dialects is common and spelling mistakes are frequent. There-
fore, the role of preprocessing is to reduce noise and prepare the feature extraction
(words, sequences of words) that will be used in the classification step.
Although the standard and the dialectal Arabic are commonly used in social
networks, the morphological differences between this forms of Arabic indicates
that each form require its own preprocessing. Arabic language has many complex
particularities such as vocabulary richness and morphological compactness. Most
words in Arabic are derived from a root with three or four characters [6]. For instance
with the root ( - to know), new words are created:
The compact aspect of Arabic results from the attachment and concatenation
capabilities for verbs, nouns and even particles. Besides the MSA, the MDA is used in
social networks frequently. The major aspect of MDA is spelling variation, therefore
with the lack of uniform rules in writing various styles can be used to express the
same phrase.
this methodology; it aims to provide clean data with less noise, so that the supervised
classification step achieve good performance [7].
Otherwise, Features construction is an intermediate step from the unstructured
textual data towards the structured feature vectors. In this study, we use two types
of configurations for features construction: the n-grams extraction scheme [7], and
feature-weighting scheme.
Polarity annotation of Facebook comments is a difficult task, since these comments
lack of polarity indicator like those used in product reviews (rating with stars or
a score). Furthermore, the sentiment expressed in Facebook comments may covers
serval topics besides the topic of interest [8]. Polarity annotation is even more compli-
cated when dealing with Moroccan Arabic comments because lexical resources dedi-
cated for this language are still insufficient. Thus, the use a human annotation was
indispensable.
The last step in the proposed methodology is the supervised classification; at this
level, the dataset is split randomly into two subsets, the learning subset is allocated to
train the supervised models, while the second subset is retained in order to evaluate
the performance.
Stemming. The task of stemming aims to group the different varieties of a word into
a single stem. Two categories of stemmers are developed for the Arabic language;
the root stemmer tries to extract the base root of the word, which is composed from
three to four characters in Arabic. [9]. Otherwise, the light stemmer remove prefixes
and suffixes of the word and retain the stem only [10]. In this work, we try to test
three different Arabic stemmers, Root stemmer, light stemmer and a dialectal light
stemmer. For root stemming we use ISRI [11], this root-extraction stemmer for Arabic
700 M. Maghfour and A. Elouardighi
is similar to the Khoja stemmer [12] but without a root dictionary. For Arabic light
stemming we try Motaz light stemmer [13] that actually modifies the ISRI Arabic
Stemmer to perform light stemming for Arabic words. For dialectal stemming, we
have developed and implemented our own new stemmer through which we have tried
to deal with the Moroccan dialectal Arabic complexities. Therefore, the dialectal
stemmer provides two stemming options: MSA or MDA, where the MDA stemming
option incorporates almost all the MSA operations, moreover it include also the
specific tasks to process the Moroccan dialect.
In this study, we have collected a corpus composed from about 10,000 Facebook
comments that were published in standard Arabic and Moroccan Arabic. The topics
of these comments were various, since they concerned economics, politics, sports
etc. A group of experts was in charge of the polarity labeling; this task resulted in
7106 negative comments and 2939 positive comments.
Several elements affect the performance in the sentiment analysis, especially the
language of the studied comments, their length and their topics [8]. For this raison,
it is important to carry out many experimentations on feature construction schemes
so that to achieve the best performance.
We have tested four text representations by combining feature-weighing tech-
niques: TF (term frequency) and TF-IDF (term frequency-inverse document
frequency) with feature extraction techniques: unigram and unigrams + bigrams.
Then, we have applied three classifiers that are already known for their robustness
in sentiment analysis [17–19], namely SVM (Support vector machines) with a linear
kernel, naive bayes and maximum entropy. For each of the four feature construction
schemes we divide the dataset into two subsets, 75% of the observations for the
702 M. Maghfour and A. Elouardighi
training set and 25% for the test set. We have applied a 10 fold cross validation
alongside with the parameter optimization to train the predictive models. The test
set, which is composed from unseen data, served us to evaluate the quality of these
models. The F-score was selected as an evaluation metric. We have used python
programming language for text preprocessing and scikit-learn library for polarity
classification.
Before presenting the results of this work, we briefly describe the steps of the iterative
approach followed for this purpose. Based on Facebook Arabic comments, four
features construction schemes were executed, and then at the classification level three
classifiers were used: Maximum entropy, Naive Bayes and SVM. For evaluation, the
F-score metric was calculated. The obtained results are presented Table 1. This results
show that the dialectal stemmer that we have developed outperforms both the root
stemmer and the light stemmer, regardless of the classifiers or the feature construction
scheme. Furthermore with this stemmer the best contribution was obtained with Naive
bayes and TF-Bigram + bigram scheme (F-score = 84.10%).
In scheme 3, the effect of stop words removal was tested. The results show signif-
icant effect of this task. Capturing negation in a sentiment analysis study was very
important because negation words have the capability of changing the expressed
meaning entirely.
The contribution of this task was observed through the results of preprocessing
scheme 4. The effect of negation task was positive in most cases with ME an NB
(expect the TF/unigram + bigram configuration). Synonym dictionary was built
in order to deal with redundancy and reduce further the sparsity of the features
vector. The results obtained with preprocessing scheme 5 were relevant. The best
performance obtained with this task concern ME with F-score = 84.94%, indeed this
result was the best in overall experimentations. Scheme 6 represents the contribution
of all proposed processing tasks; stemming, stop word, negation and dictionary.
The combined scheme enhances both NB and SVM to their maximum perfor-
mance. On one hand, the best F-score of NB (F-score = 84.53%) was achieved
under this scheme with TF/Unigram. On the other hand the best result of SVM (F-
score = 84.01%) was obtained with the configuration TF-IDF/Unigram + bigram.
Our study reveals similar results because the contribution of the Arabic root stemmer
that we have tested outperform the light stemmer, however, we have outlined that
by considering the dialectal issues in light stemming we have obtained even higher
score. Shoukry et rafea [2] carried out a SA study on Arabic tweets in Egyptian
dialect, they have tested their own stemmer that addresses Egyptian dialect issues
against an MSA light stemmer, they found better results with their implemented
dialectal stemmer. Furthermore, with stop words removal their study performance
increased by only 0.1%, which was similar to our findings. Similar works, based on
testing the effect of preprocessing on sentiment analysis, show similarities to our
Toward Improving Arabic Text Preprocessing … 703
study in term of results. In [3] duawri and El-Orfali have tested the effect of Arabic
light and root stemming on SA of politics dataset, they found that root stemming was
more helpful than light stemming.
6 Conclusion
This work was focused on the effect of processing Arabic text on SA. At first, we
have described the methodology of collecting and preprocessing Arabic Facebook
comments to perform supervised classification on sentiments. Then we have tested
704 M. Maghfour and A. Elouardighi
References
1. Pozzi F, Fersini E, Messina E, Liu B (2016) Sentiment analysis in social networks. Elsevier
Science
2. Shoukry A, Rafea A (2012) Preprocessing egyptian dialect tweets for sentiment mining. In:
The fourth workshop on computational approaches to arabic script-based languages, pp 47–56
3. Duwairi R, El-Orfali M (2014) A study of the effects of preprocessing strategies on sentiment
analysis for arabic text. J Inf Sci 40(4):501–513
4. Duwairi RM, Marji R, Sha’ban N, Rushaidat S (2014) Sentiment analysis in arabic tweets.
In: 2014 5th international conference on information and communication systems (ICICS), pp
1–6. IEEE
5. Guellil I, Adeel A, Azouaou F, Hussain A (2018) Sentialg: automated corpus annotation for
algerian sentiment analysis. In: International conference on brain inspired cognitive systems,
pp 557–567. Springer
6. Habash NY (2010) Introduction to arabic natural language processing. Synth Lect Hum Lang
Technol 3(1):1–187
7. Liu B (2012) Sentiment analysis and opinion mining. Synth Lect Hum Lang Technol 5(1):1–167
8. Pang B, Lee L et al (2008) Opinion mining and sentiment analysis. Found Trends Inf Retrieval
2(1–2):1–135
9. Duwairi R, Al-Refai M, Khasawneh N (2007) Stemming versus light stemming as feature
selection techniques for arabic text categorization. In: 2007 4th international conference on
Innovations in information technology, IIT 2007, pp 446–450. IEEE
10. Al-Anzi FS, AbuZeina D (2015) Stemming impact on arabic text categorization performance:
a survey. In: 2015 5th international conference on information & communication technology
and accessibility (ICTA), pp 1–7. IEEE
11. Taghva K, Elkhoury R, Coombs J (2005) Arabic stemming without a root dictionary. In:
International conference on information technology: coding and computing, ITCC 2005, vol
1, pp 152–157. IEEE
12. Khoja S, Garside R (1999) Stemming arabic text. Lancaster, UK, Computing Department,
Lancaster University
13. Saad MK, Ashour W (2010) Arabic morphological tools for text mining. Corpora 18:19
14. Larkey LS, Ballesteros L, Connell ME (2007) Light stemming for arabic information retrieval.
In: Arabic computational morphology, pp 221–243. Springer
Toward Improving Arabic Text Preprocessing … 705
15. Liu H, Setiono R (1995) Chi2: feature selection and discretization of numeric at tributes. In:
Proceedings, seventh international conference on tools with artificial intelligence, pp 388–391.
IEEE
16. Alwakid G, Osman T, Hughes-Roberts T (2017) Challenges in sentiment analysis for arabic
social networks. Procedia Comput Sci 117:89–100
17. Pang B, Lee L, Vaithyanathan S (2002) Thumbs up?: sentiment classification using machine
learning techniques. In: Proceedings of the ACL-02 conference on Empirical methods in natural
language processing, vol 10, pp 79–86. Association for Computational Linguistics
18. Dave K, Lawrence S, Pennock DM (2003) Mining the peanut gallery: opinion extraction and
semantic classification of product reviews. In: Proceedings of the 12th international conference
on World Wide Web, pp 519–528. ACM
19. Go A, Bhayani R, Huang L (2009) Twitter sentiment classification using distant supervision.
CS224N Project Report, Stanford 1(12)
IIMDR: Intelligence Integration Model
for Document Retrieval
Abstract With the rise of digital documents on the internet, the demand for accurate
search results has been a challenging task and many notable algorithms have been
used to cater to such queries. However, with the increasingly fast-paced discoveries
in machine learning and artificial intelligence, many of the algorithms have been
outdated and no longer used. It is necessary to display only the relevant content that
the user queried for. In this paper, IIMDR: Intelligence Integration Model for Docu-
ment Retrieval is proposed. Jaccard Similarity is used to calculate word similarity
in search result documents. Experiments using the proposed model on RCV dataset
show that the proposed framework results in improved performance by providing
more relevant documents to the user query. An accuracy of 86.78% on the RCV1
dataset has been achieved with this method.
1 Introduction
The accessibility of advanced records over the web is expanded and there is a require-
ment for compelling and snappy techniques to recover and sort out the computerized
archives. Since information is scattered universally and is sloppy, it is a provoking
assignment to create viable strategies that can produce top-notch highlights in these
archives. It is important to diminish the hole between clients’ inquiry and recovered
outcomes. In this paper, IIMDR, an Intelligence Integration Model for Document
Retrieval is proposed. Word similarity is computed using Jaccard Similarity present
in search results documents. With the rise of digital documents on the internet, the
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 707
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_64
708 S. Aditya et al.
demand for accurate search results has been a challenging task and many notable
algorithms have been used to cater to such queries. However, with the increasingly
fast-paced discoveries in machine learning and artificial intelligence, many of the
algorithms have been outdated and no longer used.
Motivation: The need for an improved and modern algorithm to identify the most
relevant documents. Modern algorithms such as Latent Semantic Analysis and
random forest classification were used in this paper which has improved the effi-
ciency and speed of the result drastically. The use of Jaccard Similarity has also
proved to produce efficient results.
2 Related Work
Landauer et al. [1] demonstrated the use of LSA and its usefulness in the anal-
ysis. Foltz [2] described the behavior of LSA and explained how semantic similarity
for text-based research works. Kostoff et al. [3] used to discover text using phrase
frequency analysis and phrase co-occurrence analysis. However, this technique didn’t
prove to be useful in the data set that was chosen for this paper as it didn’t deliver the
results that were expected. Ramya et al. [4] used Latent Semantic Analysis which
delivered very promising results. They used a DRDLC framework that worked on
the dataset RCV1 where the documents were ranked and accessed. Grineva et al. [5]
presented a novel method wherein key term extraction from text documents is done
using the properties of graphs. They found that the terms related to main topics tend
to bunch up as densely interconnected subgraphs. Pushpa et al. [6] proposed in their
paper how to adequately find applicable web administrations. They suggested that
word-level semantics is applied utilizing grouping dependent on the NPMI esteem.
IIMDR: Intelligence Integration Model … 709
3 Proposed Architecture
depicts the architecture of the proposed IIMDR framework. The steps involved are
pre-processing for removing unwanted data and reduce time complexity, Synoni-
mization and SPARQL Endpoint based enrichment to enhance the class labels of the
user query for better classification, classification of the documents to avoid comparing
irrelevant documents, and concept similarity computation to further reduce the clas-
sified documents. Reuters Corpus Volume-1 (RCV-1) is an unlabeled dataset, to
increase the performance of the recommender system it has been labeled using a
real-world domain ontology-based labeling.
3.1 Pre-processing
The input query given by the user will contain words that do not help in finding the
documents and do not help in class labeling of the user query. There are three steps
in this process. First is tokenization, since the query is in sentences, it is split into
words called tokens. The second step is Stop Word removal, words that are present
in the stop words dictionary was removed as they are used to give the sentence a
structure not meaning. The third step is stemming from this reduces the token size
to their respective root word by dropping unnecessary characters usually a suffix.
Various words in English have the same meaning, so we found the synonyms of
the tokens from the preprocessed user query. These words will be class labels for
the query which was used for the classification of the dataset using random forest
classification. To obtain the class labels of the query, WordNet 3.0 was used. Then
SPARQL Endpoint based enrichment was done to further increase the number of
class labels for document classification.
Dataset required to find the relevant document was further reduced using random
forest classification [20]. Random forest was used because it does not need any
feature scaling and provided high accuracy when comparing with other algorithms.
s
Gini = 1 − (Pi)2 (1)
i=1
IIMDR: Intelligence Integration Model … 711
In Eq. (1), Pi represents the frequency of the class observed in the dataset and s
represents the number of classes present in the dataset. Gini impurity was computed
by the random forest classifier at each branch accordingly and the classification was
done.
The classified dataset was further reduced using Concept similarity [19]. A similarity
score was assigned to each document. After scoring the documents were arranged in
descending order of the score as they are documents similar to the user query and in
these set of documents top thirty percent were taken and sent into the recommender
system.
The main aim is to retrieve the relevant document from the curated documents that
the user wants by taking the query from them. The architecture is split into three
phases Data Pre-processing, Pattern identification, and concept extraction.
The Reuter Corpus Volume 1 (RCV1) dataset utilized in this paper contains news
story website pages in XML design. These XML contains extra information such as
webpage author information, time and place of the report, and other metadata infor-
mation which can increase the computation time as they are unnecessary information
that does not contribute to the document search and increase space complexity to
store this information. Steps in Data preprocessing include XML creation (custom)
and Stop word removal and Unique term identification
Since RCV1 contains stories and news in XML design there is a necessity to prepro-
cess this because it contains unnecessary data which would lead to an increase in
time and space complexity while retrieving the relevant documents. To avoid this and
to make the framework exceptionally responsive, data that are not useful in search
activities are removed to make the XML data light. These custom XML data is then
stored in a data repository.
712 S. Aditya et al.
The common words from the dataset that does not contribute to the concept of the
question and not useful in finding the relevant documents. We take each word from
the document and check if it’s a part of the Stop word dictionary or not. Similarly,
we check whether the word is part of the unique words or not. Perform this for every
single document.
(ii) LSA (Latent semantic analysis) For concept extraction, LSA (Latent Semantic
Analysis) [22] was used, LSA performs SVD on the TF-IDF vectors for dimen-
sionality reduction. LSA itself is an unsupervised way of uncovering synonyms in a
collection of documents. LSA was used instead of other algorithms because to fill in
the “Semantic Gap” in the dimensionality reduction, by performing this we obtained
the topics of each document and stored them in the repository.
Patterns extracted from the previous phase was used to find the most similar document
to that of the user query. To find the most similar document pattern to that of the
query, Jaccard similarity [23] was implemented. The documents were ranked based
on the similarity score and the document with the highest score was the relevant
document for the given user query.
Jaccard similarity formula:
|A ∩ B|
J (A, B) = (2)
|A ∪ B|
In Eq. (2), A and B are two documents whose Jaccard similarity is to be found.
We find the ratio of common terms in A and B to the union of terms in A and B.
IIMDR: Intelligence Integration Model … 713
5.1 Implementation
The higher configuration systems may provide better results. User Query was
obtained first, then it was preprocessed where it has gone through tokenization, stop
word removal, and stemming. Next to find the documents with similar class labels
synonymizing was done to get the class labels of the query. After obtaining class
labels random forest classification was done to obtain the relevant documents from
which the top 30% of the documents were taken by assigning the documents simi-
larity score by concept similarity. This curated dataset was preprocessed to remove
the XML tags and stop words. TF-IDF was performed on the pre-processed dataset
and then LSA was performed where the concept of each document was extracted.
By doing so we have obtained the pattern of the documents. The user query’s pattern
is also obtained, and the relevant document was also found out by Jaccard similarity
the document with the highest score was chosen. Table 1 depicts the algorithm for
document retrieval in the proposed approach.
6 Performance Evaluation
LSA can be easily implemented and used. Many practical applications use LSA. LSA
is capable of assuring consistent results. It works well on diverse datasets. Further,
LSA seems to give promising results, especially in the document retrieval domain.
SPARQL was used as the query language in this experiment. As we know that
SPARQL is not constrained within one database it has been proved to be very effective
in the experiment we have performed. Cosine similarity and Jaccard similarity are
a measure of lexical similarity and the performance varies based on various factors.
However, the dataset used Jaccard similarity proved to be more efficient and gave a
more precise result.
Table 2 shows the various comparisons of various frameworks. DRDLC though
proved to be one of the precise frameworks IIMDR had performed better with the
dataset used in this experiment. Note how that LDA and LSA has performed in
combination with Jaccard. LSA has also performed more efficiently when combined
with Trigram as compared to LDA. Figure 2 depicts the FDR vs the number of
714 S. Aditya et al.
Table 1 Algorithm for Document Retrieval from the given input user query
Input: Reuters Corpus Volume 1, User query.
Output: Relevant document that user wants.
Begin
Step 1: Let each document in the data set be Di= (Ti , Ci) where Ti is text of the
document and the Ci is class labels.Let user query be Ur.
Step 2: The objective function is F(D ,Ur)=Di .Where D is total documents and Di
relevant document to the user query.
Step 3: Pre-processing of the user query.
Step 3.1: Tokenization
Step 3.2: stop word removal
Step 3.3: stemming
Step 4: Synonmization using WordNet3.0 and SPARQL Endpoint based enrichment
Step 5: Classification of the dataset using random forest algorithm to find the set of
documents similar to the class label
Step 6: Pick the top 30 percent of the documents based on the descending order of
their similarity score using concept similarity
Step 7: The documents chosen is then pre-processed
Step 7.1: creating custom XML
Step 7.2: stop word removal and unique term identification
Step 8: Pattern Identification and concept Extraction of the curated documents.
Step 8.1: TF-IDF vectorizing
Step 8.2: LSA
Step 9: The relevant document is obtained by arranging the Jaccard similarity score of
the documents with user query in descending order and choosing the first one.
End
Fig. 2 Comparison of FDR value against Number of recommendations for various frameworks
7 Conclusions
The proposed IIMDR has been modeled for recommending documents using a
hybridized approach. The approach successfully amalgamates LSI, Jaccard simi-
larity and Random Forest has been incorporated for classification. The approach uses
SPARQL endpoints for knowledge enrichment which makes it a knowledge-driven
paradigm. The hybridization enhances the outcome of the proposed IIMDR and
makes it better than the baseline models. IIMDR yields an overall accuracy of 86.78%
which makes IIMDR the best in class approach for recommending documents.
716 S. Aditya et al.
References
1. Landauer TK, Foltz PW, Laham D (2009) An introduction to latent semantic analysis
2. Foltz PW (1996) Latent semantic analysis for text-based research. Behav Res Methods Instrum
Comput 28:197–202
3. Kostoff RN, Stump JA, Johnson D, Murday JS, Lau CG, Tolles WM (2006) The structure and
infrastructure of the global nanotechnology literature. J Nanopart Res 8:301321
4. Ramya RS, Ganesh Singh T, Sejal D, Venugopal KR (2008) DRDLC: discovering relevant
documents using latent dirichlet allocation and cosine similarity. In: Conference: the 2018 VII
international conference
5. Grineva M, Grinev M, Lizorkin D (2009) Extracting key terms from noisy and multi-theme
documents. In: WWW 2009 proceedings of the 18th international conference on World Wide
Web
6. Pushpa CN, Deepak G, Kumar A, Thriveni J, Venugopal KR (2020) Improving web service
discovery by hybridization of ontology focused concept clustering and interface semantics.
IEEE. 978-1-7281-6828-9/20
7. Niwattanakul S, Singthongchai J, Naenudorn E, Wanapu S (2013) Using of Jaccard coefficient
for keywords similarity. In: 2013 Proceedings of the international multiconference of engineers
and computer scientists, IMECS 2013, vol I, 13–15 March 2013, Hong Kong
8. Deepak G, Santhanavijayan A (2020) OntoBestFit: a best-fit occurrence estimation strategy
for RDF driven faceted semantic search. Comput Commun 160:284–298
9. Pushpa CN, Deepak G, Kumar A, Thriveni J, Venugopal KR (2020) OntoDisco: improving web
service discovery by hybridization of ontology focused concept clustering and interface seman-
tics. In: 2020 IEEE international conference on electronics, computing and communication
technologies (CONECCT), pp 1–5. IEEE
10. Kumar A, Deepak G, Santhanavijayan A (2020) HeTOnto: a novel approach for concep-
tualization, modeling, visualization, and formalization of domain centric ontologies for heat
transfer. In: 2020 IEEE international conference on electronics, computing and communication
technologies (CONECCT), pp 1–6. IEEE
11. Deepak Gerard, Kasaraneni Dheera (2019) OntoCommerce: an ontology focused semantic
framework for personalised product recommendation for user targeted e-commerce. Int J
Comput Aided Eng Technol 11(4–5):449–466
12. Gulzar Z., Anny Leema A, Deepak G (2018) PCRS: personalized course recommender system
based on hybrid approach. Procedia Comput Sci 125:518–524
13. Deepak G, Teja V, Santhanavijayan A (2020) A novel firefly driven scheme for resume parsing
and matching based on entity linking paradigm. J Discrete Math Sci Crypt 23(1):157–165
14. Haribabu S, Kumar PS, Padhy S, Deepak G, Santhanavijayan A, Kumar N (2019) A novel
approach for ontology focused inter- domain personalized search based on semantic set
expansion. In: 2019 fifteenth international conference on information processing (ICINPRO),
Bengaluru, India, pp 1–5. https://doi.org/10.1109/icinpro47689.2019.9092155
15. Deepak G, Kumar N, Bharadwaj GV, Santhanavijayan A (2019) OntoQuest: an ontological
strategy for automatic question generation for e-assessment using static and dynamic knowl-
edge. In: 2019 fifteenth international conference on information processing (ICINPRO), pp
1–6. IEEE
16. Santhanavijayan AD, Naresh Kumar D, Deepak G (2021) A semantic-aware strategy for auto-
matic speech recognition incorporating deep learning models. In: Intelligent system design, pp
247–254. Springer, Singapore
17. Deepak, G, et al (2019) Design and evaluation of conceptual ontologies for electrochemistry as
a domain. In: 2019 IEEE international WIE conference on electrical and computer engineering
(WIECON-ECE). IEEE
18. Deepak G, Priyadarshini JS (2018) Personalized and enhanced hybridized semantic algorithm
for web image retrieval incorporating ontology classification, strategic query expansion, and
content-based analysis. Comput Electr Eng 72:14–25
IIMDR: Intelligence Integration Model … 717
19. Zheng C, Hu Z, Shen YP (2009) Concept similarity computation based on ontology. In: 2009
international conference on machine learning and cybernetics, vol 1, pp 165–169. IEEE
20. Strobl C, Boulesteix A, Augustin T (2007) “Unbiased split selection for classification trees
based on the Gini index” (PDF). Comput Stat Data Anal 52:483–501
21. Robertson S (2004) Understanding inverse document frequency: on theoretical arguments for
IDF. J Doc. 60(5):503–520. https://doi.org/10.1108/00220410410560582
22. Landauer TK, Foltz PW, Laham D (1998) Introduction to latent semantic analysis. Discourse
Process 25:259–284
23. Levandowsky Michael, Winter David (1971) Distance between sets. Nature 234(5):34–35.
https://doi.org/10.1038/234034a
Effective Application of Information
Technology Tools for Real-Time Project
Management
Abstract Information technology (IT) is evolving rapidly and benefits a wide variety
of industries to break conventional communication barriers. Recent advancements
in management information systems have enabled project-based organizations to
monitor and control their projects more effectively. Although the role of IT tools
(ITT) has been explored in previous research, the importance of each type of tool
in achieving higher effectiveness in real-time management of projects has been less
studied. This study aims to explore the significance of each type of IT tools for
this purpose. Thus, a survey of principal contracting organizations in the construc-
tion industry was carried out, which indicated that reporting tools, integrated plat-
forms, analytical tools, knowledge management tools, and archiving tools are key
ITT groups that should be applied collaboratively. Although reporting tools were
perceived as the most important group from the experts’ view, the correlation analysis
also revealed that all five groups are interrelated and should be applied coherently
to achieve satisfactory outcomes. This study explains the application of ITTs for
real-time project management and directs further research on this topic.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 719
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_65
720 M. Ershadi et al.
1 Introduction
The evolution of information technology tools has drastically enhanced the agility of
operations and the precision of management decisions. However, previous research
has identified the construction sector as a laggard in taking advantage of recent
technological advancement in this field [1]. Regarding the multi-player and risky
nature of the construction business, enterprises seek more robust tools to tackle
operational and functional complexities inherent in their activities. Discrepancies of
actual execution and baseline plans have led researchers to introduce the concept
of real-time project management in which automated monitoring technologies are
applied, project performance data is processed in real-time, and the plan is optimized
accordingly [2]. To achieve this target in construction projects, contractors need to
improve their understanding of the effective IT tools (ITT) to take full advantage of
recent IT evolutions for more effective project delivery.
Real-time action requires extensive application of advanced systems to improve
the integration of project planning and control processes and automate controls.
Updating the schedule based on actual performance is performed intermittently in
the traditional project control. While with the introduction of the Internet of Things
(IoT) and intelligent machines, it is possible to obtain real-time performance data
from construction sites to be analyzed and compared with baselines instantly. It
is useful in large projects that many tasks are being executed simultaneously and
minor daily deviations would aggregate and lead to major delays [1]. Besides, it
is not limited to time management but includes keeping the record of information,
lessons learned, and reports to be used for continuous improvement of processes.
It is not possible without the effective application of a variety of ITTs. Thus, this
study contributes to a better understanding of potential ITTs that can be used for this
purpose.
Prior studies highlighted the application of IT tools and systems in business activ-
ities. Akinsola et al. developed a process model based on a set of integrated ITTs
for enhancing the productivity of operations on construction sites [3]. In another
study, Hui Yap et al. discussed how ITT can be applied in managing knowledge
acquired from undertaking business activities to build, maintain, and advance knowl-
edge assets. They also discussed the challenges and barriers to implementing knowl-
edge management (KM) tools in a sample of construction organizations [4]. A
group of studies specifically focused on exploring the benefits of applying ITT for
construction organizations. For example, Love and Irani examined the approaches
used by such enterprises to evaluate and justify their investments in information
technology infrastructure, and the advantages that they have experienced due to IT
implementation.
In recent construction research, ITT is often being discussed in light of building
information modeling (BIM) that integrates technologies involved in the digital repre-
sentation of the features of a facility. For instance, Hartmann et al. explained the
BIM implementation strategies in practical contexts and advocated a “technology
push” implementation approach to align workflows with the functionality of this
Effective Application of Information Technology Tools … 721
system [5]. Besides, Cleveland explores emerging ITT such as portable virtual reality,
digital paper, and computer automatic virtual environment to be applied in support
of construction engineering practices [6]. Notwithstanding these studies, there is
a paucity of research on the application of ITT for real-time project management.
Such research provides practitioners with guidelines on the best approach to lever-
aging information technology for boosting the effectiveness of project monitoring
and control. The purpose of the present paper is to bridge this gap by identifying
main types of ITT, explores their significance from a project management (PM)
perspective, and discuss optimal ways to their implementation.
2 Theoretical Background
Information technology tools are those related to the analysis and transmit of stored
information. Different kinds of ITTs serve distinct areas depending on the context.
From the construction project management perspective, ITTs bring accurate real-time
data from construction sites to project leaders in support of their informed decision-
making within a timeframe. A variety of ITTs has been developed and introduced
for different applications in various construction subsectors [3]. The result of the
literature review revealed that five categories of ITTs have been introduced for project
management. These categories include (1) archiving tools, (2) analytical tools, (3)
reporting tools, (4) KM tools, and (5) integrated platforms (Table 1).
These five groups of ITTs improve PM processes and provide a suitable platform
for better integration that facilitates the establishment of real-time control systems.
Archiving tools allow systematic storing of project artifacts both during the stages
of execution and at the delivery stage. The tools in this category have the same
features as document management systems (DMS) that help organizations to locate
3 Methods
4 Results
The one-sample t-test was applied in this study to examine whether the ITTs identified
earlier were significant to be effectively applied for real-time project management.
This statistical test was carried out in SPSS 25 software (Table 2). The significance
level for all five variables is less than 0.05, which means that according to experts’
viewpoint they are essential tools for real-time management of construction projects.
This t-test revealed that the mean rating of the variables is significantly higher than
the cut-point 3 considering 95% confidence level.
The ITTs were ranked based on the ratings given by industry experts so that their
importance level can be identified. The literature asserts that the relative importance
index (RII) can be applied rather than the mean value to provide insight into the
Effective Application of Information Technology Tools … 725
ranking of variables based on a normalized index. This index is worked out using
formula 1 in which w stands for the ratings of ITTs given by experts based on the
Likert scale, A represents the highest weight to be given to each question (5 in the
Likert scale), and N stands for the total number of experts who participated in the
survey [24].
W
RI I = (1)
A×N
The ranking of the variables based on RII revealed that reporting tools are
perceived as the most important group of ITTs, which play an important role as
a coordinator in collecting data from user interfaces, obtain results of the data anal-
ysis from analytical tools, and bring them to the attention of decision-makers. The
integrated platform followed by analytical tools was identified as the second most
important ITT. It is noteworthy that the RII score of these variables is close to each
other and higher than 0.75 which confirms their dependencies (Table 4).
5 Discussion
6 Conclusion
The intense interactions among frequent stakeholders that are involved in construc-
tion projects have led contractors to adopt advanced IT tools for enhancing commu-
nications. This paper aims to examine the importance of ITTs and discuss their appli-
cation for a real-time project management system in this demanding environment.
Five groups of ITTs were examined in this empirical study. Project reporting tools
were perceived as the most effective categories of tools that enable higher visibility
and transparency of information between key stakeholders. Effective monitoring and
control of construction activities require collecting precise data on team performance,
risks, resource productivity, progress, needs, and issues to be addressed promptly.
Analytical tools analyze collected data to be used for planning, control, and manage-
ment of projects. The analytical ITTs that are employed for analyzing data from
multiple projects consider interdependencies among their activities, resources, and
constraints to help optimally plan their resources. This group of ITTs helps to adopt
proper strategies for allocating plant and labor resources to a portfolio of projects.
The outputs of analytical software should help managers to identify the root cause
of issues from a systematic perspective.
This study examined the effectiveness of ITTs in storing and retrieving documents
and knowledge obtained from previous and current projects. Archiving tools act
as a document management system that keeps the record of different versions of
documents for the instant access of project teams. On the other hand, knowledge
management ITTs provide a basis for better transmit of lessons learned both at an
inter-project and intra-project level. The study asserted that these two groups of
tools perform better in conjunction with each other to link knowledge and relevant
digital documents as an approach to integrated record management. Four groups
of ITTs were found to be complementary to each other and interrelated in their
inputs/outputs, which indicates that they need to be consolidated in a single integrated
platform. Integrated platforms help to centralize data and facilitate the exchange and
interoperability of project information. Such platforms nurture standardization and
728 M. Ershadi et al.
consistency in the adoption of ITTs for centralized management and real-time control
of several projects in support of their on-target delivery.
References
1. Fulford R, Standing C (2014) Construction industry productivity and the potential for
collaborative practice. Int J Proj Manag 32(2):315–326
2. Isaac S (2019) What does “real-time” project control really mean? In: European conference on
computing in construction: Chania, Crete, Greece
3. Akinsola A, Dawood N, Hobbs B (2000) Construction planning process improvement using
information technology tools. In: Proceedings of CIT, pp 40–51
4. Yap JBH, Lock A (2017) Analysing the benefits, techniques, tools and challenges of knowledge
management practices in the Malaysian construction SMEs. J Eng Des Technol 15(6):803–825
5. Hartmann T et al (2012) Aligning building information model tools and construction
management methods. Autom Constr 22:605–613
6. Cleveland A Jr (2011) Emerging tools to enable construction engineering. J Constr Eng Manag
137(10):836–842
7. Latimer D, Hendrickson C (2003) Digital archival of construction project information. Nist
Special Publication SP, pp 59–64
8. Azhar S, Brown J, Farooqui R (2009) BIM-based sustainability analysis: an evaluation of
building performance analysis software. In: Proceedings of the 45th ASC annual conference,
vol 1, no 4, pp 276–292
9. Siew RY, Balatbat MC, Carmichael DG (2013) A review of building/infrastructure sustain-
ability reporting tools (SRTs). Smart Sustain Built Environ 2(2):106–139
10. Perez-Araos A et al (2007) Designing a knowledge management tool to support knowledge
sharing networks. J Manufact Technol Manag 18(2):153–168
11. Saka AB, Chan DW (2020) Profound barriers to building information modelling (BIM) adop-
tion in construction small and medium-sized enterprises (SMEs). Constr Innov 20(2):261–284
12. Blackburn DA, et al (2009) Construction imaging and archiving method, system and program.
Google Patents
13. Subramani T, Jabasingh DS, Jayalakshmi J (2014) Analysis of cost controlling in construction
industries by earned value method using Primavera. Int J Eng Res Appl 4(6):145–153
14. Hu W, Panthi K (2020) Project status reporting system (PSRS) for pipeline relocation programs.
Built Env Proj Asset Manag 10(5):693–708
15. Wu S et al (2004) A project knowledge management tool for the construction industry. Int J IT
Archit Eng Constr 2(2):79–90
16. Xu G et al (2018) Cloud asset-enabled integrated IoT platform for lean prefabricated
construction. Autom Constr 93:123–134
17. Zou PX et al (2017) Cloud-based safety information and communication system in infrastruc-
ture construction. Saf Sci 98:50–69
18. Yang J-B, Kao C-K, Lee Y-Y (2006) System requirement analysis of a construction delay
analysis system. In: Proceeding of 23rd international symposium on automation and robotics
in construction, pp 102–106
19. Shen X, Marks E (2016) Near-miss information visualization tool in BIM for construction
safety. J Constr Eng Manag 142(4):04015100
20. Chu SKW (2008) TWiki for knowledge building and management. Online Inf Rev 32(6):745–
758
21. Hsiao DW et al (2009) An integrated platform of collaborative project management and silicon
intellectual property management for IC design industry. Inf Sci 179(15):2576–2590
22. Adhau S, Mittal M, Mittal A (2013) A multi-agent system for decentralized multi-project
scheduling with resource transfers. Int J Prod Econ 146(2):646–661
Effective Application of Information Technology Tools … 729
23. Liaw S-S, Chen G-D, Huang H-M (2008) Users’ attitudes toward Web-based collaborative
learning systems for knowledge management. Comput Educ 50(3):950–961
24. Chan DW, Kumaraswamy MM (1997) A comparative study of causes of time overruns in Hong
Kong construction projects. Int J Proj Manag 15(1):55–63
25. Schuh G, et al (2016) Next generation hardware development: Framework for a tailorable devel-
opment method. In: 2016 Portland international conference on management of engineering and
technology (PICMET). IEEE
26. Chen W-N, Zhang J (2012) Ant colony optimization for software project scheduling and staffing
with an event-based scheduler. IEEE Trans Softw Eng 39(1):1–17
27. Francis A, Miresco E (2014) Case studies for the planning and monitoring of unit-and
fixed-price contracts using project scheduling software. In: Computing in civil and building
engineering, pp 1021–1028
28. Dang T, Bargstädt H-J (2016) 4D relationships: the missing link in 4D scheduling. J Constr
Eng Manag 142(2):04015072
HCODF: Hybrid Cognitive Ontology
Driven Framework for Socially Relevant
News Validation
Abstract Nowadays everyone gets news from social media and most often it is
very much difficult to validate its credibility. Social media platforms like twitter and
WhatsApp are working on this for a long time, lots of research work has been done
by many people but still an effective technique hasn’t been found out. This paper
proposes a method to address this problem, with the use of a hybrid NLP model that
uses an Event Ontology and LSTM to validate whether a news is authentic or not. The
news dataset that has been used here are the tweets crawled and the news headlines,
where the semantic similarity between the news and tweets on specific events are
found out, which is later used to classify the tweets whether it is fake or not. This
experiment’s performance is evaluated and it is compared with baseline models and
it was found that the proposed approach is superior in terms of performance which
recorded an F-Measure and False Negative Rate of 96.14% and 0.03 respectively.
1 Introduction
In this modern world, anyone can publish any information, on the Web although that
is acceptable up to certain extent, but taking it too far such as posting wrong infor-
mation online in order to trigger and manipulate people, causing panic in the society,
spreading rumors about a person to tarnish their public image because of personal
vengeance can lead to everlasting and ever ending repercussions. We cannot put an
end to this problem once for all but by constantly developing and improving existing
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 731
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_66
732 V. Adithya et al.
technique’s that address this issue, impact of this problem can be reduced. Machine
Learning was the primary strategy incorporated in most Fake News Detection systems
[1].
Is all the news that we tend to come across on the social media are true? Social
media is the fastest mode of sharing and spreading information that can reach a mass
of people within minutes. In that case spreading fake or unauthentic news via social
media causes a lot of problems. Nowadays many people have come to a conclusion
that all news that comes in social media are false or very old messages that is been
circulated again. It is true in most of the cases, users are confused whether the news
is true or not.
Motivation. Social media platforms like WhatsApp and twitter are working on this
problem for a long time but still an effective technique hasn’t been found for it. The
solution to this problem can be found by using Natural Language Processing (NLP)
which is used to process text data. WhatsApp forward message that keeps circulating
in a circle, where no information of the source is included. In contrast to WhatsApp,
the twitter tweets have the information of the source and does not contain a large
number of sentences, at the same time reach a large number of audiences. Thus, by
utilizing the advantage, the tweets can be crawled easily by filtering the retweets and
can be easily processed.
Organization. The latter portion of the paper is divided into the following sections.
Section 2 addresses briefly, the relevant research done related to this topic. Section 3
explains the planned system architecture. Section 4 consists mainly of the imple-
mentation and performance evaluation. Section 5 consists of conclusion drawn from
the experiment.
2 Related Work
Jain et al. [1] have put forth an NLP and machine learning model for fake-news
detection, which uses SVM and Naïve Bias with an accuracy of 93.6%. Seo et al.
[2] have proposed a methodology that uses question and answering system, which
uses modified CNN deep learning model and they also say that their model archives
high accuracy when compared to statistical models. The model focuses on data
HCODF: Hybrid Cognitive Ontology Driven Framework … 733
augmentation batch size control and model modification. Samantaray et al. [3] have
divided the proposed model into two parts, text analysis and performance evaluation,
text similarity approach by using N-Grams, TF-IDF and cosine similarity, corpus-
based similarity and term-based similarity matchings have been used. Bahada et al.
[4] have used Bi-directional LSTM-Recurrent Neural Network model applied on two
publicly present unstructured news articles for fake news detection and say that it
produces high accuracy when compared with CNN model, vanilla-RNN model and
unidirectional-LSTM model for the same task. Ajao et al. [5] have used a hybrid
CNN and RNN models for fake news detection on twitter. The proposed architecture
issued an accuracy of 82%. Authenticity has been validated based on the contents
of the messages, semantic attributes, linguistic characteristics similar to that of the
fake news. In [6–16] several ontology models and applications have been discussed
in support of the proposed methodology.
The architecture of the proposed fake new detection HCODF system is as shown in
the Fig. 1. The proposed HCODF system consists of 3 phases, the 1st phase is slot
filling using LSTM [17] event classifier for the news database, 2nd phase is Event
Ontology creation for the twitter data and the 3rd phase is comparing both with
Jaccard Similarity.
The main purpose of this HCODF system is to find out the similarity between the
Event Ontology that is created for the tweets present in the twitter dataset and the
slot filling, which is done for the news article for the particular event by the LSTM
event classifier model which is used to classify whether it is authentic or not. The
first phase involves pre-processing of the raw news headlines data, by using a bag
of words model, the headlines are tokenized into a set so that the sentence structure
is lost and then these sets are combined using a union operation and lowercase
conversion, number to text conversion, punctuations removal, stop word removal,
word stemming, lemmatization is done. Then the pre-processed data is ready for the
Term Frequency-Inverse Document Frequency TF-IDF.
TF-IDF is a technique, generally used for Information retrieval, and is used to
calculate the significance of every word in the document and rank them accordingly
so that the relevancy of each word in the document is found out which can be used
as features for the dataset. The Eq. (1) represents the formulation of TF-IDF as a
product of Term Frequency and Inverse document Frequency.
N
IDF(t) = (3)
DF
N
IDF(t) = log( ) (4)
DF + 1
N
TF − IDF(t, d) = TF(t, d) × log (5)
DF + 1
Equation (3), (4) shows how Eq. (5) the final TF-IDF equation is derived.
After TF-IDF is encompassed, the data is used to train the LSTM such that it
generates keywords for the events that is used for slot filling. The LSTM is also
kind of RNN architecture that has feedback connections. It can process the whole
sequence of the data. Normally, a LSTM unit consists of a cell, which is used to
store values across time periods and three gates namely, input, output and forget
gates. These gates are utilized to control the flow of information into the cell and
vice versa. LSTM’s are used for classification and for prediction using time-series
data. In a similar manner, the LSTM used here is used to classify the type of the
event that the news headlines is talking about which is used later for filling the slots.
The twitter tweets are crawled for each event that has been tweeted between a
particular duration of time in English and the retweets are filtered and is saved with
its timestamp. The tweets are crawled using a tweet crawling api called as twint. The
HCODF: Hybrid Cognitive Ontology Driven Framework … 735
The proposed HCODF system’s architecture was designed and successfully imple-
mented on Windows 10 operating system. The implementation has been performed
using Intel Core i5 8th Gen processor and 8 GB RAM. NLTK, Re, num2words and
sklearn are the python NLP tool libraries that has been used for pre-processing data
and performing TF-IDF. Keras, a python’s deep learning API library running on top
of TensorFlow has been utilized to implement and train the LSTM Event Classifier
model. Web Protege an open-source online Ontology editor created by the Protege
has been used to create and visualize ontologies for all the events.
The experiment was performed on the top news headlines of India dataset, for
the past 19 years, a standard dataset published in Kaggle of which the past 5 years
data has been utilized and some data set has been crawled using news crawler api
for some specific events and both are combined. The tweets were crawled for events
with a twitter crawler api called as twint by using the event’s respective hash tag
as keywords and the retweets are filtered and the data is stored with its time stamp.
The Precision, Recall, Accuracy, F-Measure and False Negative Rate are calculated,
and used as performance evaluation metrics. The Accuracy, F-Measure, and False
Negative Rate are calculated and represented as Table 1 and Table 2 using the Eqs. (7),
(8), (9), (10), (11) respectively.
Table 1 Performance of the Proposed HCODF system for some Viral Events
Viral News Precision % Recall % Accuracy % F-Measure % FNR
Pulwama Terror Attack 87.69 90.12 88.91 88.89 0.10
Demonetization in India 92.63 96.78 94.71 94.66 0.03
Cyclone Vardha 90.14 92.63 91.39 91.37 0.07
Cyclone Winston 89.43 94.43 91.93 91.86 0.06
Amazon Forest Wildfire 91.69 96.47 94.08 94.02 0.04
Precision + Recall
Accuracy % = (9)
2
2(Precision × Recall)
F − Measure % = (10)
(Precision + Recall)
The experiment was trained on several events. For this experimentation purpose, 5
events are considered. The Table 1 shows the performance evaluation for 5 individual
events namely, Pulwama Terror Attack, Demonetization in India, Cyclone Vardha,
Cyclone Winston, Amazon Forest Wildfire. It can be observed that the Precision,
Recall, Accuracy, F-Measure And False Negative Rate for Pulwama attack is 87.6,
90.12, 88.91, 88.89% and 0.10 respectively, for Demonetization in India it is 92.63,
96.78, 94.71, 94.66% and 0.03 respectively, for Cyclone Vardha it is noted to be
90.14, 92.63, 91.39, 91.37% and 0.07 respectively, for Cyclone Winston it is 89.43,
94.43, 91.93, 91.86% and 0.06 respectively and for Amazon Forest Wildfire it is
91.69, 96.47, 94.08, 94.02% and 0.04 respectively.
For this implementation, 3 existing approaches are considered as baseline as
shown in Table 2. The first approach proposed by Jain et al. [1] uses traditional NLP
with SVM and Naive Bayes and our HCODF model’s performance increase is when
compared to it is 2.85% in precision, 3.39% in recall, 3.12% in Accuracy, 3.11%
in F-measure and 0.03 in False Negative Rate. The second approach proposed by
Seo et al. [2] uses modified convolution neural networks to perform the same task
and our HCODF model’s performance-increase when compared to it is 9.46% in
Precision, 10% in Recall, 9.73% in Accuracy, 9.71% in F-measure and 0.10 in False
Negative. The Third approach proposed by Samantaray et al. [3] a hybrid of TF-IDF,
N gram vectors and cosine similarity. HCODF model’s performance-increase when
compared to it is 4.75% in Precision, 4.93% in Recall, 4.44% in Accuracy, 4.43% in
F-Measure and 0.04 in False Negative Rate.
The superiority of the HCODF system when compared to the baseline model
in terms of performance is because it uses a hybridization both LSTM and an
Event Ontology model. The HCODF model also uses Jaccard similarity instead
of cosine similarity when compared to the third approach, which is also responsible
for increase in its performance. Since the HCODF uses LSTM when compared to
the first approach that uses traditional NLP and the second approach that uses CNN,
the increase is performance is justified. The Event Ontology model adds secondary
knowledge to the system when compared to using similarity function without an
Ontology. Which is the primary reason for improvement in performance.
738 V. Adithya et al.
5 Conclusions
There is a necessity to validate the authenticity of the news that is published in social
media because the news spreads to a greater number of people when compared to any
other mode. This paper discusses an effective method for the same. This proposed
HCODF system can perform accurate predictions when compared to others because
of the Ontology driven framework, where an Event Ontology model is combined
with LSTM and used as a hybrid. Based on this, it can be concluded that in future,
ensuring techniques with hybrid approaches combined with Ontology driven frame
works can produce better results for the same purpose fulfilment. An overall F-
Measure of 96.14% with a very low FDR of 0.03 was obtained. As a future work
the proposed HCODF can be experimented with other fusion models of machine
learning, deep learning, and other similarity models.
References
1. Jain A, Shakya A, Khatter H, Gupta AK (2019) A smart system for fake news detection using
machine learning. In: 2019 2nd international conference on issues and challenges in intelligent
computing techniques (ICICT)
2. Seo Y, Seo D, Jeong C-S (2019) FaNDeR: fake news detection model using media reliability. In
Proceedings of TENCON 2018 - 2018 IEEE region 10 conference, pp 1834–1838). [8650350]
(IEEE Region 10 annual international conference, proceedings/TENCON, vol 2018-October.
Institute of Electrical and Electronics Engineers Inc. https://doi.org/10.1109/TENCON.2018.
8650350
3. Samantaray SD, Jodhani G (2019) Fake news detection using text similarity approach. Int J
Sci Res (IJSR) 8(1):1126–1132
4. Bahada P, Saxenaa P, Kamalb R (2019) Fake news detection using bi-directional LSTM-
recurrent neural network. In: International conference on recent trends in advanced computing
2019, Icrtac 2019
5. Oluwaseun A, Deepayan B, Shahrzad Z (2018) Fake news identification on Twitter with hybrid
CNN and RNN models. SMSociety, Copenhagen, Denmark, July 2018
6. Gerard D, Santhanavijayan A (2020) OntoBestFit: a best-fit occurrence estimation strategy for
RDF driven faceted semantic search. Comput Commun 160:284–298
7. Pushpa CN, Deepak G, Kumar A, Thriveni J, Venugopal KR (2020) OntoDisco: improving web
service discovery by hybridization of ontology focused concept clustering and interface seman-
tics. In: 2020 IEEE international conference on electronics, computing and communication
technologies (CONECCT). IEEE, pp 1–5
8. Kumar A, Deepak G, Santhanavijayan A (2020) HeTOnto: a novel approach for concep-
tualization, modeling, visualization, and formalization of domain centric ontologies for heat
transfer. In: 2020 IEEE international conference on electronics, computing and communication
technologies (CONECCT). IEEE, pp. 1–6
9. Deepak G, Kasaraneni D (2019) OntoCommerce: an ontology focused semantic framework
for personalised product recommendation for user targeted e-commerce. Int J Comput Aided
Eng Technol 11(4–5):449–466
10. Gulzar ZA, Leema A, Deepak G (2018) PCRS: personalized course recommender system based
on hybrid approach. Procedia Comput Sci 125:518–524
11. Deepak G, Teja V, Santhanavijayan A (2020) A novel firefly driven scheme for resume parsing
and matching based on entity linking paradigm. J Discrete Math Sci Crypto 23(1):157–165
HCODF: Hybrid Cognitive Ontology Driven Framework … 739
12. Haribabu S, Sai Kumar PS, Padhy S, Deepak G, Santhanavijayan A, Kumar DN (2019) A
novel approach for ontology focused inter-domain personalized search based on semantic set
expansion. In: 2019 Fifteenth International Conference on Information Processing (ICINPRO),
Bengaluru, India, 2019, pp 1–5. https://doi.org/10.1109/icinpro47689.2019.9092155
13. Deepak G, Naresh Kumar G, Sai Yashaswea Bharadwaj VSN, Santhanavijayan A (2019) Onto-
Quest: an ontological strategy for automatic question generation for e-assessment using static
and dynamic knowledge. In: 2019 fifteenth international conference on information processing
(ICINPRO). IEEE, pp 1–6
14. Santhanavijayan A, Naresh Kumar D, Deepak G (2021) A semantic-aware strategy for auto-
matic speech recognition incorporating deep learning models. In: Intelligent system design.
Springer, Singapore, pp 247–254
15. Gerard D et al (2019) Design and evaluation of conceptual ontologies for electrochemistry as a
domain. In: 2019 IEEE international WIE conference on electrical and computer engineering
(WIECON-ECE). IEEE (2019)
16. Deepak G, Priyadarshini JS (2018) Personalized and enhanced hybridized semantic algorithm
for web image retrieval incorporating ontology classification, strategic query expansion, and
content-based analysis. Comput Electr Eng 72:14–25
17. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9:1735–1780.
https://doi.org/10.1162/neco.1997.9.8.1735
Understanding Customer Adoption
of E-banking
1 Introduction
The banking sector is one of the sectors that has taken a strong interest in digitization.
Indeed, the speed of growth of technology has more influence on banks sector than
on others [1]. In addition to that, the international economic crisis and the resulting
crisis of confidence in the markets has affected several financial institutions. It has
pushed banks to try to maintain customers relationship in different ways [2].
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 741
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_67
742 J. Chama et al.
2 Literature Review
E-banking is one of the services offered by banks to provide quick access to users
through electronic channels. With the development of computer technology, elec-
tronic banking services are increasingly accepted by customers who strongly support
new technologies [5]. They can be used to attract new customers because they allow
the transmission of information in real time and provide a convenient channel for
communication between the bank and its customers. Because of this, electronic
banking services encourage customers to perform financial operations more easily
and efficiently through electronic channels.
Asad, Mohajerani and Nourseresh [6] has conducted research that attempts to
explain the factors affecting customer satisfaction in electronic banking and to estab-
lish priority factors. According to these authors, several factors affect customer
satisfaction, such as usefulness, trust, security, web design, information content and
responsiveness through customer support services that must be provided.
In their turn, Ling, Fern, Boon and Huat [5] conducted a study in Malaysia
through a questionnaire survey of 200 customers who have experience using elec-
tronic banking services. Their aim was to examine the factors that influence customer
satisfaction with these services. The five factors identified by the authors are service
quality that is formed when the customer makes a comparison between the expected
service and the actual service experience, web design and content related to the
configuration of colors and graphics, security and confidentiality which is a form of
protection to ensure customers safety, convenience that corresponds to usefulness
and speed associated with network connection failures and internet banking website
browsing time.
Understanding Customer Adoption of E-banking 743
Aliyu, Rosmain and Takala [4] also developed a conceptual model based on several
previous studies related to electronic banking services, behavioral factors, banking
applications and customer service delivery. The authors conducted a hypothetico-
deductive study by administering an online questionnaire to 283 clients in Malaysia.
Their results revealed that security, cost and convenience are the major drivers of
customer service delivery in electronic banking. The study also demonstrated the
significant link between customer satisfaction and the use of electronic banking
services.
The aim of the study conducted by Al-Somali, Gholami and Clegg [7] was to
answer a set of questions, to know what factors influence the trend of customers
to accept and use online banking services as their primary banking channel? What
is the relative importance of these factors? and what is the nature and strength of
the relationship between them? The research structure was developed on the basis
of the Technology acceptance Model by integrating new control variables and then
tested with 202 clients in Saudi Arabia. The authors’ conceptual model suggests that
awareness of online banking services and their benefits is a critical component of
E-banking adoption as much as customers may not be familiar with them because
it is a new experience for many of them. Trust is a bigger issue in online banking
than in traditional branch. It is very important in the online banking industry because
by nature transactions contain sensitive information for customers. According to the
authors, social influence also plays a key role in the acceptance and behavior of
E-Banking users, it corresponds to the subjective standards defined by the opinions
of consumers entourage. The study also highlights the presence of the resistance to
change factor, that is the shift from traditional banking to online banking for some
customers who are not ready to change their current ways of operating. Of course,
demographic factors such as age, gender or level of education cannot be overlooked.
According to this study, they also have an impact on customer attitudes and behaviors
towards online banking services.
In Germany, Jünger and Mietzner [3] looked at the factors that determine the deci-
sion of households to switch from traditional banks to new financial technologies.
The authors’ results show the dominance of three main dimensions which are trust,
transparency and financial expertise. For Jünger and Mietzner, these factors signif-
icantly influence household decisions to stay or switch from traditional banking
services to new financial technologies.
the ease felt when using a new technology will encourage its adoption. In fact, this
is the second significantly important determinant in technology adoption within the
Technology Acceptance Model. It refers to the degree to which the potential user
expects the use of the system to be effortless [8]. So, it is about the perceived ease
when using technology. Ease of use of electronic channels should therefore be sought
as a necessary determinant for the use of E-banking services. In addition, bad design
and aesthetics can prevent the user from revisiting the site [6].
3.3 Trust
Trust is a key element of success that is present in the banking sector in both traditional
and electronic channels. Previous research has shown that trust is central to build long-
term relationships between customers and companies, especially in risky electronic
banking environments [12]. Indeed, lack of trust is one of the most frequently reasons
cited by consumers who do not use electronic banking services in numerous studies
around the world [12–18].
Social influence is also an E-banking adoption factor which is very present in many
research studies. According to Baabdullah, Alalwan, Rana, Kizgin and Patil [19],
it is precisely the extent to which an individual perceives that important people
believe he should apply the new system. According to the authors, the preferences
and values of those around us, family members, friends, neighbors and other users
of the technology, tend to change the views and perceptions of E-banking users.
Understanding Customer Adoption of E-banking 745
In terms of cost, several authors have stated that to encourage consumers to use new
technologies, their adoption must be reasonably priced compared to the prices of
other alternatives. The notion of price value present in the literature is defined as the
cognitive compromise of consumers between the perceived benefits of E-banking
and the monetary cost of adopting this service. With a high level of price value,
customers become more enthusiastic about continuing to use E-banking technology
[13].
Behavioral intention is a predictor of technology adoption that has been widely and
repeatedly reported to have an important role in shaping actual use and adoption of
new systems [13]. This is a factor present in many technology adoption models such
as TAM or UTAUT and UTAUT2. Defining behavioral intention is quite simply to
answer the following question: is the client initially willing to accept and use a new
technology?
4 Recommendations
Based on the results of previous work which enabled us to identify the main adoption
factors mentioned above, we can say that today’s customers have higher requirements
towards E-banking. According to Hamidi and Safareeyeh [20], they are more knowl-
edgeable, want to get more products and more valuable information. To this end, we
can make some recommendations to practitioners in the field of E-Banking that they
can implement to support the adoption of their electronic channels.
First of all, we strongly recommend increasing the ease perceived by customers
through clear and comprehensive electronic support as well as the establishment of
customer support. Then, banking institutions must ensure the strengthening of their
electronic banking security system in order to create and maintain customer’s trust. It
is also advisable to minimize costs when customers use E-banking services in order
to increase the price value and thereby increase its use.
5 Conclusion
Due to advanced technology and innovation in the financial and banking sectors,
electronic banking services have become more familiar to customers of traditional
746 J. Chama et al.
banks [5]. Indeed, many customers are opting more and more for the novelty of
E-Banking, given the many advantages that it offers.
In order to increase the use of this electronic channel, this research has enabled us
to highlight the most common E-banking adoption factors in the literature, namely
ease of use, facilitating conditions, perceived usefulness, performance expectancy,
trust, social influence, cost, price value and behavioral intention.
However, by removing the social component of business exchange which is the
heart of relationship marketing, information and communication technologies can,
in some cases, degrade the relationship between the customer and the company [21].
In the case of E-Banking for example, customers can easily consult other websites
that offer similar services or products and this means that customer loyalty would
no longer be an easy task [19]. We can therefore ask ourselves if these technologies
really contribute to the development of the relationship with the customer and the
differentiation between banking brands in an environment where all competitors use
the same technologies.
References
13. Alalwan AA, Dwivedi YK, Rana NP (2017) Factors influencing adoption of mobile banking
by Jordanian bank customers: extending UTAUT2 with trust. Int J Inf Manage 37:99–110
14. Lee KC, Chung N (2009) Understanding factors affecting trust in and satisfaction with mobile
banking in Korea: a modified DeLone and McLean’s model perspective. Interact Comput
21:385–392
15. Masrek MN, Mohamed IS, Daud NM, Omar N (2014) Technology trust and mobile banking
satisfaction: a case of Malaysian consumers. Procedia Soc Behav Sci 129:53–58. 2nd
International Conference on Innovation, Management and Technology Research
16. Mostafa R (2015) Investigating the role of trust in mobile banking acceptance, in: ideas in
marketing: finding the new and polishing the old, pp 834–842
17. Sharma SK, Sharma M (2019) Examining the role of trust and quality dimensions in the actual
usage of mobile banking services: an empirical investigation. Int J Inf Manage 44:65–75
18. Merhi M, Hone K, Tarhini A (2019) A cross-cultural study of the intention to use mobile
banking between Lebanese and British consumers: extending UTAUT2 with security, privacy
and trust. Technol Soc 59:101151
19. Baabdullah AM, Alalwan AA, Rana NP, Kizgin H, Patil P (2019) Consumer use of mobile
banking (M-Banking) in Saudi Arabia: towards an integrated model. Int J Inf Manage 44:38–52
20. Hamidi H, Safareeyeh M (2019) A model to analyze the effect of mobile banking adoption on
customer interaction and satisfaction: a case study of m-banking in Iran. Telematics Inform
38:166–181
21. Ailli S (2016) Impact des Technologies de l’Information et de la Communication sur le
marketing relationnel: une évaluation des opérateurs des banques commerciales au Maroc.
Revue Marocaine de recherche en management et marketing 19
Amalgamation of Novel Objective
Function and Multi-sink Solution
for a Reliable RPL in High Traffic
Monitoring Applications
Abstract Recent monitoring applications are based on Low power and Lossy
networks (LLN), standardized by Internet Engineering Task Force (IETF). These
kinds of networks require a trustworthy routing protocol for nodes aggregation and
good QoS achievement. RPL gains popularity compared to other routing protocols
for its flexibility via its objective function. Default objective functions implemented
in the RPL core are based on a single metric. Consequently, the routing protocol
can’t cope with different constraints and show some issues in high traffic applica-
tions. For that, we proposed in our paper a cross approach between combined metrics
linearly and lexically for the novel objective function supported by multiple sinks
solution for network monitoring. The proposal, evaluated under the COOJA emulator
against Standard-RPL, showed a reliability improvement by increasing the PDR and
throughput, decreasing considerably the power consumption and giving stability to
the network through low number of DIO control packets.
1 Introduction
Data monitoring using the wireless sensor networks, considered as a great connec-
tivity potential, are tiny sensor devices [1] based networks with constraint and limited
resources [2]. Consequently, there is a need for appropriate resources management
and utilization to cope with different applications as smart grids [3], smart cities [4],
industries [5] and healthcare [6]. Thereby, in order to ensure performances efficiency
of those networks, the researchers have been interested mostly in routing protocols.
In this context, the most commonly used is RPL designed by Internet Engineering
Task Force (IETF) [7]. The Ipv6 Routing Protocol for Low power and lossy networks
(RPL) is based on IPv6 and uses the IEEE802.15.4 at the PHY and MAC layers [8].
Packets based on IPv6 cannot fit in the IEEE 802.15.4 protocol. For this reason,
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 749
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_68
750 A. Eloudrhiri Hassani et al.
6LowPAN, IPv6 Low Power Personal Area Network layer [9], acts as an adaptation
between MAC and network layers. With RPL, the paths are constructed once the
network is initialized. The nodes aim to set up destination-oriented directed acyclic
graph (DODAG), a tree routing topology, using four principal ICMPv6 messages:
DODAG Information Object (DIO) holds information that enable nodes to know
the instance, configuration and preferred parent, it managed by the trickle algo-
rithm and necessary for Multiple Points to Point (MP2P) and Point-to-Point (P2P)
communications [10]. DODAG Information Solicitation (DIS) as a DIO request from
neighbors. DODAG Advertisement Object packets (DAO) used to collect downward
topology informations and DAO-ACK as a response to a DAO message. The aim of
DODAG topology is to steer data packets to one or multiple sinks. Routing paths
are created using an objective function (OF). In the core of RPL, two OF’s are
proposed namely the minimum rank with hysteresis objective function (MRHOF)
[11] based on expected transmission count as a routing metric and OF0 [12], which is
based on hop count. However, the two OF tend to minimize the cost of their metrics
which causes some paths optimization issues due to taking into account a single
constraint. Furthermore, in high traffic applications case, routing to a single sink
may show limits in terms of reliability and ensuring the nodes lifetime. Considering
these drawbacks, we have proposed an amalgamation between novel designed objec-
tive function, which uses novel strategies for metrics combination, and multiple sinks
solution for the network monitoring. The rest of this paper is organized as follow. In
Sect. 2, we present the related works. In Sect. 3, we present the proposed solution
for RPL improvement. In Sect. 4, we report the performance results and discussion,
finally a conclusion is given in Sect. 5.
2 Related Works
In this section, we present some related researches to our work that improve the
objective functions for RPL. In [13], San Martin et al. proposed an extended metric
based on the expected transmission count called Sigma-ETX which is the standard
deviation value of ETX by number of hops to sink. This technique avoids the long
hops issue that causes bottlenecks in high network densities. Also ELT metric, i.e.
Expected Lifetime, was designed by Iova et al. in [14] that consist of the node
remaining lifetime. It is injected in RPL based on multipath, which leads to enhance
the reliability despite an additional delay induced. Minimizing the delay of receiving
packets by the sink was considered by Gonizzi et al. in [15] using the AVGDELAY
as a routing metric. This technique showed a decrease in terms of end-to-end delay
but don’t address the network reliability at all. The fuzzy logic approach was consid-
ered by Bhandari et al. in [16]. Indeed, the authors proposed COOF, i.e. a novel
objective function context-oriented objective function, that considers two designed
metrics respectively remaining energy (RE) and queue fluctuation index for smart
cities applications. Using the same approach in [17], Lamaazi et al. fuzzy designed
EC-OF. It combines ETX and node energy consumption while hop count as node
Amalgamation of Novel Objective Function … 751
redirection metric. Results showed that the proposal improves the RPL performances
in terms of packet delivery ratio, latency, convergence, power consumption including
the network lifetime against MRHOF. Sankar et al. in [18] proposed Fuzzy Logic
Based Energy Aware for RPL objective function based on residual node energy,
expected transmission count ETX and traffic load. It is used for calculating the rank
increase parameter for nodes rank. The proposal shows a increase of PDR and life-
time. Congestion problem was addressed in [19] by proposing Congestion-Aware
Objective Function. Al-Kashoash et al. have used the buffer occupancy in conjunc-
tion with the expected transmission count metric. The proposed objective function
provides high packet delivery ratio and throughput while it decreases the energy
consumption. In this context, Eloudrhiri Hassani et al. In [20] designed Forwarding
Traffic Consciousness objective function, to cope with congestion due to the high
number of forwarded packets, which combines hop count, rssi and a newly designed
FTM metric. The proposed method showed a packet delivery ratio increase respec-
tively with 2 and 11% in low and high traffics, considerably reduces the power
consumption with approximately 47% as well as it achieves a good balance of traffic
managed by the relay nodes. The additive and lexical approaches were also the base
of several objective function optimizations. Indeed, Eloudrhiri Hassani et al. in [21]
proposed a new objective function based on additive combination of node and link
metrics respectively ETX and Energy consumption along the routing paths. They
proposed two way of combination, i.e. with weights and non weighted methods.
Results show that the WCMOF and NWCMOF can increase the reliability, maxi-
mize the network life time and reduce the parent changes. Smart grid application
requirements were considered by Nassar et al. in [22]. The authors designed OFQS
objective function based on power state metric, expected transmission count and
delay. The stability was considered by the hysteresis concept as in MRHOF. Results
showed improvement in terms of end-to-end delay, network lifetime and packet
delivery ratio. In [23], Al-Kashoash et al. have been interested in the congestion of
the path caused by the buffer nodes occupancy. For that purpose, Congestion-Aware
Objective Function were be proposed which consider ETX metric at a low data rate
while the buffer occupancy is considered at a high data rate.
3 Proposed Contribution
The quality of services offered by a routing protocol for IoT based on wireless sensor
networks is one of the substantial design concerns. Besides, its multiple requirements
meeting allow it to be widely used for different applications. As mentioned before, the
RPL objective function is mainly responsible for satisfying the application prerequi-
sites according to the constraint metrics used in the preferred parents selection. In a
standard manner, RPL employs an objective function with one metric to minimize.
It leads to cope with limited constraints, which certainly inflict poor performances in
term of various requirements. Furthermore, the RPL protocol hardly copes with high
traffic applications since there is one sink that monitors the network. To overcome
752 A. Eloudrhiri Hassani et al.
these challenges, we have proposed a novel objective function for routing data with
RPL supported by the multiple sinks solution as shown in Fig. 1 in order to increase
the network quality of services.
Number of Hops: is the only metric announced by neighboring nodes. For that
purpose, we have modified the structure of the DIO control message precisely in the
metric container part. Indeed, at reception of DIO message by a node i, it process
the hop count metric following Eq. 2.
Amalgamation of Novel Objective Function … 753
Minhcincrement cp = s
HC (i) = (2)
HC (n) + MinhcIncrement cp = s
Where rssi_reg(i) is a 8 bits register that gives the signal strength between node i
and its neighbor. The antenna variation offset is represented with adjustment equal
to 45 provided from the CC2420 radio datasheet. Since the Rssi metric gives values
between 0 and −110 dBm Respectively for good and bad signal strength, this metric
must be maximized. To keep the objective function convergence, we worked with its
absolute value to follow the minimizing rule as other combined metric.
During the process of parent selection, nodes are based on ranks calculated by the
objective function. Neighbors with lower rank are elected by the concerned node to
be its preferred parent. Unlike standard objective functions based on a single metric,
our proposal uses a set of metrics combined for rank processing and making the
routing decision. Indeed, at a DIO reception from a neighbor, the node calculates the
number of hops and the rssi at the MAC layer. The combination of these metrics is
done linearly with different weights following Eq. 4.
Where Rank (N) is the rank of the neighbor candidate to be the preferred parent
through node i could forward packets to sink. At this stage, once the node has calcu-
lated its rank based on the dio sender, if it is the first round for network establishment,
the node chooses it as the preferred parent. otherwise, the lexical approach is acti-
vated by a comparison firstly between the etx metrics of old and candidate rank, if it
754 A. Eloudrhiri Hassani et al.
is higher; the candidate parent is automatically discarded. else, if the candidate parent
rank is lower too than the preferred parent rank, then the candidate parent is hold.
finally, if the condition for the parent change is fulfilled, the node updates its rank
and the hop count metric in the dio message, then broadcast a new one to advertise
neighbors that there is a change in the routing path. Table 1 show the proposed work
algorithm implemented in the RPL core.
The first part of reliability investigation, we evaluate the Packet Delivery Ratio (PDR)
defined as the quotient of the packets sent by all the nodes of the network during the
period simulated by the number of packets received at the sink node. Figure 2 show
that our proposal with single sink provides an average PDR increase with 5.3, 4, 15
and 11% for 10, 20, 30, 60 ppm respectively against Standard-RPL. Also, according
to the multiple sinks solution added to our proposed objective function, an increase
about 30 and 37% respectively with 2 and 4 sinks is provided. These results can be
explained by the fact that, with multiple sinks, the congestion can be avoided and the
collaboration between destinations for the network monitoring can divide a single
network into several mini-networks which allows better management of the nodes
then better reliability.
The second part of reliability investigation, we evaluate the throughput, considered
as an after effect of the packet delivery ratio. Indeed, it is defined as the total received
packet by the sink or all the sinks divided by the total simulation time. Figure 3 show
that our proposal with single sink provides an average throughput increase with 6%,
3.5%, 18% and 13% for 10, 20, 30, 60 ppm respectively against Standard-RPL. Also,
according to the multiple sinks solution added to our proposed objective function,
an increase about 28% and 36% respectively with 2 and 4 sinks is provided.
Fig. 2 Packet delivery ratio for different number of sent packets per minute
756 A. Eloudrhiri Hassani et al.
Fig. 4 Power consumption of nodes for different number of sent packets per minute
Amalgamation of Novel Objective Function … 757
Fig. 5 DIO control packets for different number of sent packets per minute
Also, the multiple sinks solution can provide a decrease about 14 and 38 respectively
with 2 and 4 sinks against one sink solution.
The after effect of network stability is provided by the number of DIO control packets.
Indeed, as shown in Fig. 5, our proposal with single and multiple sinks can provide a
low number of DIO packets compared to Standard-RPL due to the optimized paths
selected using combined metrics and stability given by the mini network management
by each sink.
5 Conclusion
References
Abstract Wireless Sensor Network (WSN) technology plays an important role in the
application of smart irrigation by monitoring environmental soil conditions, which
helps determine the amount of water required for an agricultural field. These sensors
are exposed to harsh environmental conditions over long period of time and became
prone to failure. These failures may lead the base station to take the wrong decision.
This Paper presents a centralized approach toward detecting and isolating sensor fault
in WSN using analytical redundancy which to achieve water saving and guarantee
a good crop quality. Simulation results are proved the effectiveness of the proposed
method.
1 Introduction
In the last decades, the world has experienced a lot changes such as climate change
with its effect on precipitation, and the population growth that led to an increase
in food demand. These significant changes call for the adoption of a smart irriga-
tion system in the agriculture field. Smart irrigation has been the subject of multiple
research efforts recently and several research present the architecture of smart irriga-
tion system using WSN technology [1–4]. This system did not just prove its efficiency
on the yield and water saving but also it enables the automation of control and moni-
toring the crops remotely. WSN plays an important role in this application, the sensor
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 759
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_69
760 N. Jihani et al.
nodes are scattered over a farm in order to monitor and gather data concerning agri-
cultural conditions such as temperature, humidity, rainfall, soil PH and send them
wirelessly to the base station. Based on these monitored conditions, the base station
performs the required calculations to determine whether the agricultural field needs
water and calculate the amount of water required. To achieve the desired results, WSN
should operate correctly and accurately during the agricultural season. In smart irri-
gation system, the sensor nodes are exposed to harsh environmental conditions over
long period of time. This leads to a high probability of sensor nodes malfunction
which can cause fault measurement. The latter can cause the base station to make
faulty decisions which may damage or reduce the quality of the crops and causes
water waste. So, the WSN requires continuous monitoring to verify the quality of
measurements to ensure a reliable monitoring of the field and therefore obtaining an
adequate irrigation process, and most of the proposed smart irrigation systems have
not addressed this issue.
Fault detection and isolation (FDI) approach aims to detect and isolate faults in the
engineering systems using analytical redundancy rather than hardware redundancy
which is need the extra equipment and maintenance cost. FDI concept has widely
been used in computer science and information technology for several years [5, 6].
However, the use FDI in WSN for smart irrigation is not widely investigated.
FDI approaches in WSNs can be classified into two categories: centralized FDI
and distributed FDI. The first category, requires a central node or a base station to
collect data from the nodes, then analyzes it and takes decision if the collected data
are free faults or not. While in the second category, each node autonomously detects
and isolates its sensors faults and may also detect and isolate the sensors faults of its
neighbors.
In [7], the authors solved the FDI over WSN with centralized approach while in
[8] and [9], a distributed approach is adopted for detecting and isolating faults in
WSN.
In [8] a distributed FDI algorithm is developed in which each node detects and
isolates its sensors faults and the faults of its neighbors using parity space approach.
Each node built its model based on the measurements of its sensors and the measure-
ments of its neighbors. The same thing for [9], a model is built according to the
measurements of its neighbors using neural network approach. In some way a
distributed approach proved its efficiency on network lifetime but still has a weakness
over time in detection accuracy, nodes may make their decision based on incorrect
measurements from neighbors. Plus the limited computation capacity of the nodes
cannot support complex computation as neural network. While the main tasks of the
nodes in the most applications are to monitor environmental conditions and route
them to the base station.
Wireless Sensor Network Fault Detection … 761
However, in [7] the faults are detected and isolated with a centralized way, but
the proposed method initially requires knowing the characteristics of the faults and
their impact on the characteristics of the measured signal and based on that, for each
type of sensor fault, they created a specific algorithm to detect it.
This paper presents a centralized approach for detecting and isolating fault in WSN
using an analytical redundancy approach, the parity space approach. This proposed
method does not require a model of the monitored system or to know beforehand the
behavior of the faulty component and the characteristic of the fault as used in [7],
and the centralized approach also helps keeping detection accuracy over time and
facilitates the nodes’ tasks contrary to the distributed approach [9].
The remainder of the paper is organized as follows: Sect. 2 presents the problem
statement. Section 3, the proposed approach for FDI in a WSN with a centralized
architecture. Simulation and results discusses in Sect. 4. Finally, conclusion is given
in Sect. 5.
2 Problem Statement
The framework of the proposed smart irrigation is shown in Fig. 1. A set of sensor
nodes deployed over a farm toward measure environmental parameters such as
temperature and soil moisture, and then with the collaboration of these nodes, the
data are transformed wirelessly to the base station where the data are analyzed with
sensor fault detection and isolation (SFDI) algorithm before being used to make the
decision to irrigate the farm or not.
We consider the measurement of each sensor node writing as the following
expression:
Where yi (k) is the measurement vector at node i, x(k) is a vector that represents
the target quantities, Ci is the measurement matrix at node i that indicates which
quantities are measured at node i and f i (k) is the fault vector.
The gathered measurements from sensor nodes, form a model of the network in
the gateway as below:
⎡ ⎤
y1 (k)
⎢ y (k) ⎥
⎢ 2 ⎥
⎢ ⎥
⎢ . ⎥
y(k) = ⎢ ⎥ (2)
⎢ . ⎥
⎢ ⎥
⎣ . ⎦
yi (k)
3 Residual Generation
First up, we generate a global residual based on the model of the network with the
parity space approach, by multiplying the model Eq. 2 by ω:
Where ω is a projection matrix which makes the residual independent of x(k) and
sensitive just to the occurrence of the faults and it is calculated as follow:
ωc = 0 (4)
Fault Detection. The generated residual as shown in Eq. 5, sensitive to the occur-
rence of the fault in the network. If the fault affects any sensor node, the norm of the
residual will be high than zero otherwise the norm of the residual will be equal zero.
if fi = 0 ||r || = 0 (7)
Wireless Sensor Network Fault Detection … 763
Fault Isolation. Based on the generated residual and on the parity matrix (ω) which
its columns define n distinct fault signature directions, we can isolate the faulty sensor
by multiplying transpose of each column of the parity matrix by the residual.
F Ii = ωiT ∗ r i = 1, 2, . . . n. (8)
This function measure the correlation of the residual vector with fault signature
directions. So, if the fault is detected, the largest value of F I i corresponds to the
faulty sensor [10].
4 Simulation Results
Where y1 (k), y2 (k) and y3 (k) are the measurements recorded by sensor node 1,
2 and 3 respectively.
Based on this model, an implemented algorithm in the base station generates a
global residual and analyses it as explained in Sect. 3 for detecting and isolating
sensor fault.
In this paper, we examined only a single fault. So each time we inject a fault in
one sensor node, our algorithm detects and isolates it. Below, the simulation results
demonstrates the work that we did.
Based on the measurements recorded by the WSN, our algorithm generates a
residual norm as it is shown in Fig. 3. At t = 30 s a fault of type ‘abrupt’ is injected
into sensor node 1. As can be seen from Fig. 3, the generated residual norm is sensitive
to the fault and deviates from zero at t = 30 s, which means that from this instant,
there is a fault detected in the WSN.
After fault detection in WSN by the residual norm. To isolate the faulty sensor
node, our algorithm automatically calculates the correlation of the residual vector
with each column of the parity matrix (ω) as explained in Sect. 3.
764 N. Jihani et al.
In our test, we have 3 fault signature directions defined by the column of the
parity matrix; the first, second and the third column of ω correspond to the fault
signature direction of the sensor 1, 2 and 3 respectively, the algorithm calculates the
correlation of the residual vector with each column of the parity matrix (ω) (Eq. 8)
and the largest value of F I i corresponds to the faulty sensor. As can be seen from
Fig. 4, F I 1 has the largest magnitude which means the faulty sensor is the senor 1.
Wireless Sensor Network Fault Detection … 765
At t = 30 s, the fault is injected into sensor node 2, the Fig. 5 shows that the global
residual norm of the network starts to react when the fault is present.
And the Fig. 6 shows that the F I 2 has the largest magnitude and that is mean, the
faulty sensor is sensor 2.
The fault is inserted into sensor 3 at t = 30 s and Fig. 7 shows that the residual
norm becomes different from zero from this instant.
766 N. Jihani et al.
The fault isolation process shows that F I 3 has the largest magnitude in Fig. 8. So
the faulty sensor is sensor 3.
Wireless Sensor Network Fault Detection … 767
5 Conclusion
References
1. Riquelme JL, Soto F, Suardíaz J, Sánchez P, Iborra A, Vera JA (2009) Wireless sensor networks
for precision horticulture in Southern Spain. Comput Electron Agric 68(1):25–35
2. Ojha T, Misra S, Raghuwanshi NS (2015) Wireless sensor networks for agriculture: the state-
of-the-art in practice and future challenges. Comput Electron Agric 118:66–84
3. Abd El-kader SM, El-Basioni BMM (2013) Precision farming solution in Egypt using the
wireless sensor network technology. Egyptian Inform J 14(3):221–233
4. Kassim MRM, Mat I, Harun AN (2014) Wireless Sensor Network in precision agriculture
application. In: 2014 international conference on computer, information and telecommunication
systems (CITS), pp 1–5. IEEE, July
5. Staroswiecki M, Hoblos G, Aitouche A (2004) Sensor network design for fault tolerant
estimation. Int J Adapt Control Signal Process 18(1):55–72
6. Isermann R (2011) Fault-diagnosis applications: model-based condition monitoring: actuators,
drives, machinery, plants, sensors, and fault-tolerant systems. Springer
7. Peng Y, Qiao W, Qu L, Wang J (2017) Sensor fault detection and isolation for a wireless
sensor network-based remote wind turbine condition monitoring system. IEEE Trans Ind Appl
54(2):1072–1079
8. Hao J, Kinnaert M (2017) Sensor fault detection and isolation over wireless sensor network
based on hardware redundancy. J Phys Conf Ser 783(1):012006. IOP Publishing
9. Smarsly K, Law KH (2014) Decentralized fault detection and isolation in wireless structural
health monitoring systems using analytical redundancy. Adv Eng Softw 73:1–10
10. Patton RJ, Chen J (1991) A review of parity space approaches to fault diagnosis. IFAC Proc
Vol 24(6):65–81
Multi-resolution Texture Analysis
for Osteoporosis Classification
1 Introduction
M. Mebarkia · S. Khemaissia
Laboratoire de Génie Electrique (LABGET), Larbi Tebessi University, Tebessa, Algeria
e-mail: [email protected]
A. Meraoumia (B) · L. Houam
LAMIS Laboratory, Larbi Tebessi University, Tebessa, Algeria
e-mail: [email protected]
L. Houam
e-mail: [email protected]
R. Jennane
I3MTO Laboratory, University of Orleans, Orléans, France
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 769
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_70
770 M. Mebarkia et al.
Several treatments have been shown to reduce the risk of osteoporotic fractures by
following the necessary steps that can be taken to diagnose osteoporosis and thus
prevent bone fracture [3, 4]. An early diagnosis of osteoporosis helps in preventing
bone fractures. The diagnosis of osteoporosis relies primarily on measuring BMD
(expressed as a T-score) using dual-energy X-ray absorptiometry (DXA). DXA is
considered as the reference-standard examination for BMD assessment. Unfortu-
nately, the high-cost, low-availability, and inconsistency of BMD measurements do
not make it a promising tool for the diagnosis of osteoporosis [5, 6]. In addition,
BMD does not cover the whole process of osteoporosis diagnosis, skeletal factors
and micro-architecture features are also of great significance in the early diagnosis
of osteoporosis.
In accordance with many studies [7–9], the particular texture analysis of trabec-
ular bone radiographs is a promising approach to assess and recognize this disease
effectively. Image feature information of trabecular bone has been known to be corre-
lated closely to bone density changes and therefore can be used for the evaluation
of osteoporosis. Many algorithms have been used for the extraction of features of
trabecular pattern, such as Gabor filter, wavelet transform and fractal models [10–12].
Song et al. have presented another approach which again used the analysis of bone
textures by the usage of Fisher encoding of local descriptors algorithm with SVM
for different classes [13], whereas Nasser et al. have proposed another approach to
assess osteoporosis based on the sparse autoencoder and SVM classifier [14]. Zheng
et al. have presented a concept based on a calculation of sparse representations, to
classify bone texture image patterns [15].
Recently, with the advancements of Machine learning, deep CNNs have been
recognized as effective tools for image classification [16]. For instance, Ran Su et al.
combined CNN features with hand-crafted features for osteoporosis classification
[17]. However, previous approaches relied mainly on handcrafted categorized feature
indexes [14, 18], and machine learning (ML) algorithms, such as Support Vector
Machine (SVM) and fuzzy classifiers [9]. In addition, most of the previous studies
require tedious process and manual operations, such as extensive preprocessing,
image normalization, and region of interest (ROI) extraction, which can drastically
affect the reproducibility of the classification method [17].
The purpose of the present work is to validate the improved methods known
as Three Patches Local Binary Patterns (TP-LBP), and Four Patches Local Binary
Patterns (FP-LBP), suggested in [19]. The approach is composed in four stages proce-
dure. First, a discrete wavelet transform (DWT) is performed to the original images
of bone radiographs. In second step, TP-LBP/-FP-LBP methods are performed on
all resulting DWT sub-bands at each decomposition level, to construct the feature
vectors as shown in Fig. 2. The third step, the Fisher score algorithm is applied to the
feature vector as supervised feature selection method. Finally, the SVM classifier is
used with the linear kernel for classification of two different populations composed
of osteoporotic (OP) and control subjects CS).
Multi-resolution Texture Analysis for Osteoporosis Classification 771
The rest of this paper is organized as follows: Sect. 2 presents and discusses
in details the proposed feature extraction process. Section 3 deals with the experi-
mental results in which unimodal and multimodal systems are discussed. Finally, the
conclusion and future work are presented in Sect. 4.
The performance of all pattern recognition systems depends mainly on the distinction
of feature vectors extracted from the data. In fact, the role of the feature extraction
process is to represent the images with a reduced and distinct representation, so
that the system can discriminate the different patterns within a reasonable time. In
our proposal, we used the texture analysis method with a multi-resolution analysis
technique (e.g. DWT).
1 if x ≥ τ
f (x) = (2)
0 if x < τ
772 M. Mebarkia et al.
Fig. 1 Extended local binary pattern. a Three-Path Local Binary Pattern (TP-LBP) and b Four-Path
Local Binary Pattern (FP-LBP) [19]
Where d denotes the distance between patches, C is the patch, and f is a threshold
function and τ is the threshold of comparison.
Four-Path LBP (TP-LBP). The FP-LBP is another extended method of LBP. It uses
two rings of radius r 1 and r 2 centered in the pixel, and S patches distributed around
these two rings of (w × w) size (see Fig. 1.(b)) [19]. The FP-LBP is formed by the
comparison between two center symmetric patches in the inner circle with two center
symmetric patches in the outer circle by taking one bit in each pixel into accounts
according to the similarity of the two patches. The FP-LBP operator is given by the
following equation:
S −1
FP-LBPr 1,r 2,S,w,α = 2
f (d C1,i , C2,i+αmod(S) − d(C1,i+ S , C 2,i+ S +αmod(S) ))2i (3)
i=0 2 2
The structure of the proposed feature extraction method is presented in Fig. 2. The
image is first transformed using the DWT, then the features are extracted using
Multi-resolution Texture Analysis for Osteoporosis Classification 773
the LBP algorithm and finally, the extracted information is combined into the two
system’s levels (feature extraction and feature matching levels).
The input image (Ao ) of size (H × W ) is first transformed using the DWT into
several decomposition levels, as follows:
FDWT
o
(Ao ) = Ao ∈ R H ×W
i = 1, 2, . . . . (4)
FDWT (Ai−1 ) = [Ai , Hi , Vi , Di ] ∈ R 2i × 2i
H W
i
Where F DWi
T denotes the DWT transform at i level and Ai , Hi , Vi , Di are the
resulting decomposition bands at level i. Then, distinctive features of each band
are extracted separately using the TP(FP)-LBP technique.
Where FTP(FP)-LBP denotes the TP-LBP (FP-LBP) technique and one of the resulting
decomposition bands (A: Approximation, H: Horizontal, V: Vertical, D: Diagonal).
After extraction, the system can: i) either, combine (fuse) all the extracted feature
vectors into one single vector (FF: Fusion at feature extraction level) then carry on
the classification process, ii) or, continue the classification process for each vector
alone till the end and then combine the classification results for all the vectors into
a single result (FS: Fusion at the matching score level).
Fusion at Feature Extraction Level. For its simplicity, the concatenation method
is used in this study. Let V Ai , VHi , VVi , VDi be the feature vectors of the Ai , Hi , Vi , Di
bands at level i, the combined vector is obtained as follows:
Where FCONCAT combines the vectors by placing them in front of each other.
Fusion at Matching Scores Level. Here, each band works independently, and their
results are combined to obtain a scalar score that will be used later for decision
making. In our study, the sum of weighted scores defined as follows is used:
n 1
dF = wi di , wi = n (7)
i=1 Ei 1
j=1 Ei
Where d F denotes the fused score, di is the score obtained by the subsystem i,
n the number of subsystems, Ei represents nthe system error and wi is the weight
associated with the subsystem i. Note that i=1 wi = 1 and the weights are inversely
proportional to the corresponding errors and are therefore larger for the most precise
systems.
774 M. Mebarkia et al.
3 Experimental Results
This section highlights the performance of the proposed method by being integrated
into an osteoporosis detection system. Thus, to evaluate the effectiveness of the
proposed approach, we used a dataset of 174 images acquired at the hospital of
Orleans (France) from postmenopausal women, using X-rays on the calcaneus site
[23]. The dataset comprised 87 Osteoporosis Patients OP (with fractures) and 87
Control Subjects (CS).
Several experiments are conducted within two main parts. The first part of the exper-
iments determines the optimal parameters for each feature extraction methods (TP-
LBP and FP-LBP). The second part focuses on the performance evaluation of the
proposed osteoporosis detection system. This part is also divided into two sub-parts,
the first concerns the study of the unimodal system (each DWT band is treated
separately), while the second part focuses on the effect of the data fusion on the
performance of the system (multi-modal system).
the Curve (AUC) equal to 96.83% (Accuracy (ACC) = 91.76%), 97.22% (99.18%)
and 96.07% (90.59%), respectively. Compared to TP-LBP, the system performance is
very efficient for all patch sizes with FP-LBP. Thus, under the different patch sizes, the
AUC reaches 99.02% (ACC = 97.06%), 99.97% (99.41%) and 99.57% (98.24%) for
w equal to 3, 5, 7, respectively. It can be noted that excellent performance exceeding
96.00% for both TP-LBP and FP-LBP are achieved. In conclusion, it is clear that for
each feature extraction method, the value of 5 achieves the best performance.
Fig. 3 Performance of the unimodal osteoporosis detection system. a, c AUC for TP-LBP and
FP-LBP, respectively and b, d ACC when using TP-LBP and FP-LBP, respectively
with an AUC of 88.71% (ACC = 80.02%), i.e. a degradation of 8.75% in the case of
the horizontal detail band (cH). In this configuration, the best accuracy is obtained
using the detail band (cD) with an ACC equal to 82.94% (a degradation of 9.04%),
as in Fig. 3(b). In the case of the second method (FP-LBP), the system generally
remains slightly unchanged with respect to the first method (see Fig. 3 (c) and (d). In
this case the system works well (best AUC = 94.63% with ACC = 87.65%) with the
cA band, the degradation rates in AUC and ACC are 5.34% and 11.83%, respectively.
Therefore, the obtained results in all decomposition bands are suitable. In the next
subpart, the osteoporosis detection system will be examined for all the decomposition
bands when the first and second decomposition levels (L1 and L2 ) are fused. The
fusion will be done at the feature extraction level and at the matching score level,
then a comparison of the performance of the system will be addressed for the whole
image.
classification rates. Data fusion can reduce system errors and thus improve their
efficiency. Therefore, this subpart aims to examine whether the system performance
can be improved by using a combination of information extracted from different
DWT-bands. Thus, the feature vectors extracted from these bands are combined to
make the system more efficient than a unimodal system.
Fusion at the feature extraction Level (FF) makes it possible to combine different
vectors from different processing and analysis phases to form a single feature
vector. Therefore, after decomposing the image using DWT and obtaining the four
bands, features are extracted for each band separately and then the system combines
(concatenates) the extracted vectors to get a single feature vector. Afterward, the
system continues until the final decision is made (classification). In our experiments,
all the DWT bands from each level (L1 and L2 ) are combined and the result (as AUC)
is obtained using the whole image. Results are reported in Table 2.
In fusion at matching score level (FS), the individual scores are combined to form
a single score which is then used to make the final decision. It is the most widely used
type of fusion because it can be applied to all types of systems in a simple and efficient
way. In our multimodal system, each DWT-band has its own subsystem, which means
that the overall system contains four subsystems that operate in parallel. Then the
results of the subsystems are combined for the final decision. As previously, we
also combined the DWT bands from each level of decomposition to examine the
possibility of improving the performance of the system. The results obtained are
shown in Table 2. As can be seen, it is clear that although data fusion did not improve
the classification result, the results obtained are very satisfactory in both feature
extraction methods. In the first method (TP-LBP), the system operates with an AUC
equal to 97.19% (ACC = 93.53%), while the second method (FP-LBP) remains more
efficient with an AUC = 98.69% (ACC = 94.12%). Figure 4 illustrates a comparison
between unimodal and multimodal systems.
Figure 4 shows that data fusion can give effective and promising results, especially
when considering parallel processing.
Fig. 4 Osteoporosis detection performance under one level of DWT decomposition. a Performance
comparison using TP-LBP and b Performance comparison using FP-LBP method
References
7. Park C, Took CC, Seong J-K (2018) Machine learning in biomedical engineering. Biomed Eng
Lett 8:1–3
8. Kavitha MS, Asano A, Taguchi A, Kurita T, Sanada M (2012) Diagnosis of osteoporosis from
dental panoramic radiographs using the support vector machine method in a computer-aided
system. BMC Med Imaging 12:1
9. Muthu Subash K, Pugalendhi Ganesh K, Soon-Yong P, Kyung-Hoe H, Min-Suk H, Takio K,
Akira A, Seo-Yong A, Sung-Il C (2016) Automatic detection of osteoporosis based on hybrid
genetic swarm fuzzy classifier approaches. vol 45, p 20160076
10. Houam L, Hafiane A, Boukrouche A, Lespessailles E, Jennane R (2012) Texture character-
ization using local binary pattern and wavelets. Application to bone radiographs. In: 2012
3rd International Conference on Image Processing Theory, Tools and Applications (IPTA), pp
371–376
11. Taleb-Ahmed A, Dubois P, Duquenoy E (2003) Analysis methods of CT-scan images for the
characterization of the bone texture: First results. Pattern Recogn Lett 24:1971–1982
12. Oulhaj H, Rziza M, Amine A, Toumi H, Lespessailles E, Hassouni ME, Jennane R (2017)
Anisotropic discrete dual-tree wavelet transform for improved classification of trabecular bone.
IEEE Trans Med Imaging 36:2077–2086
13. Song Y, Cai W, Zhang F, Huang H, Zhou Y, Feng DD (2015) Bone texture characterization
with fisher encoding of local descriptors. In: 2015 IEEE 12th international symposium on
biomedical imaging (ISBI), pp 5–8
14. Nasser Y, Hassouni ME, Brahim A, Toumi H, Lespessailles E, Jennane R (2017) Diagnosis
of osteoporosis disease from bone X-ray images with stacked sparse autoencoder and SVM
classifier. In: 2017 international conference on advanced technologies for signal and image
processing (ATSIP), pp 1–5
15. Zheng K, Jennane R, Makrogiannis S (2019) Ensembles of sparse classifiers for osteoporosis
characterization in digital radiographs. Med Imag Comput Aid Diag 1095024
16. Singh A, Dutta MK, Jennane R, Lespessailles E (2017) Classification of the trabecular bone
structure of osteoporotic patients using machine vision. Comput Biol Med 91:148–158
17. Su R, Liu T, Sun C, Jin Q, Jennane R, Wei L (2020) Fusing convolutional neural network
features with hand-crafted features for osteoporosis diagnoses. Neurocomputing 385:300–309
18. Oulhaj H, Rziza M, Amine A, Toumi H, Lespessailles E, Jennane R, El Hassouni M (2017)
Trabecular bone characterization using circular parametric models. Biomed Signal Process
Control 33:411–421
19. Wolf L, Hassner T, Taigman Y (2008) Descriptor based methods in the wild. In: European
conference on computer vision workshop on faces in real-life images (ECCV), Marseille,
France
20. Ojala T, Pietikäinen M, Harwood D (1996) A comparative study of texture measures with
classification based on featured distributions. Pattern Recogn 29:51–59
21. Ojala T, Pietikainen M, Maenpaa T (2002) Multiresolution gray-scale and rotation invariant
texture classification with local binary patterns. IEEE Trans Pattern Anal Mach Intell 24:971–
987
22. Houam L, Hafiane A, Boukrouche A, Lespessailles E, Jennane R (2014) One dimensional local
binary pattern for bone texture characterization. Pattern Anal Appl 17:179–193
23. Lespessailles E, Gadois C, Kousignian I, Neveu JP, Fardellone P, Kolta S, Roux C, Do-Huu JP,
Benhamou CL (2008) Clinical interest of bone texture analysis in osteoporosis: a case control
multicenter study. Osteoporosis Int 19:1019–1028
High Performance Decoding
by Combination of the Hartmann
Rudolph Decoder and Soft Decision
Decoding by Hash Techniques
Abstract Decoding algorithms are used to recover information after its transmission
through an unreliable radio channel. Soft decision decoding is very performant in
concatenation solutions that exploit many decoding stages. Therefore, we carry out
a combination of a symbol-by-symbol decoder, which is Hartmann Rudolph (HR)
algorithm, and the Soft Decision Decoder by Hash Techniques (SDHT) as a word-to-
word decoder. In this paper, we introduce a concatenation of HR Partially exploited
(PHR) and the SDHT algorithm to decode Bose-Chaudhuri-Hocquenghem (BCH)
codes. This scheme enables at first to employ Hartmann Rudolph algorithm using
only a part of the dual codewords, then the output of PHR is treated by the SDHT.
Our proposed serial concatenation guarantees very good performances although we
use very low percentage of the dual code space. As instance for the BCH(31, 21, 5)
code, the obtained good results are based only on 21.29% of the dual codewords.
1 Introduction
Recently, there has been an expanding demand for perfect and reliable digital data
transmission and backup systems. This demand has grown by the emersion of exten-
sive, high throughput data networks for the exchange, treatment, and backup of binary
information in different sectors and domains. A considerable concern of standards
organizations, which develop protocols for mobile telephony, is the check of errors
so that authentic reproduction of data can be achieved. Wherefore, error correction
codes have been developed. Error correction coding is achieved by joining redun-
dancy to the transferred message using an algorithm. Several correction codes are
implemented in various equipment like smartphones, CDs, DVDs, storage media or
packets transmitted through Internet and cellular networks.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 781
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_71
782 H. Faham et al.
There are two types of decoders employed in communication systems: soft deci-
sion and hard decision decoders. Soft decision algorithms exploit directly the received
symbols and they employ principally the Euclidian distance as a measure. While
the hard decision decoders measure is usually the Hamming distance. This type of
algorithms treat binary inputs.
The rest of the work is arranged in this way. In the second section, some decoders
are presented as cognate works. The third section presents the suggested serial
concatenation of PHR and SDHT. The fourth section gives detail about our proposed
algorithm experiments. Finally, we give a conclusion through the fifth section.
2 Related Works
SDHT [13] is a Soft decision Decoder at the base of Hash Techniques and
syndrome decoding. The originators of [13] have introduced two hard decision
decoders named HSDec [14] and HWDec [14], and Chase-HSDec decoding algo-
rithm [15], which exploits HSDec in the Chase-2 decoder with the aim of minimizing
its temporal complexity.
In [16], it was proved that off-the-shelf Deep Learning decoding algorithms can
achieve the maximum a posteriori (MAP) decoding performance for short length
codes. In [17] a performance comparison of novel Chien search and syndrome blocks
designs is given for BCH and Reed-Solomon codes.
3.1 HR Algorithm
Let us note the linear code by C(n, k, d), k is the code dimension, n is its length
and d is the
minimum distance between all its distinct vectors. Moreover, let us note
t = d−1 2
the error correcting capability of C. The Soft Decision Decoding by Hash
Techniques operates as following:
784 H. Faham et al.
HR has proposed to exploit all the dual codewords. This makes the temporal
complexity of this algorithm very high especially for codes whose of parity bits
number is relatively large. With the aim of overcoming this constraint, Nouh et al.
[18] have suggested exploiting just M dual codewords, and then the HR is used only
in some symbols of the received sequence. Therefore, the algorithm will be named
Partial Hartman-Rudolph (PHR). In [19] we have applied the same idea. The formula
(1) becomes (2).
M n 1−φl c 1 ⊕δml
cm = 0 i f j=1 i=1 ( 1+φl )
jl >0
(2)
cm = 1 other wise
The fact of reducing the used dual codewords number influences the decoding
quality. It is why we propose to reprocess the sequence returned by HR by a word-
by-word decoder, which is the SDHT algorithm.
The PHR-SDHT decoder works as follows:
High Performance Decoding by Combination of the Hartmann Rudolph Decoder … 785
With the intention of showing our suggested scheme efficacy, we provide in this
section the PHR-SDHT simulation results for some BCH codes. Some comparisons
of PHR-SDHT performances with other competitors are also given. Table 1 describes
the employed experiment parameters.
The performances of decoding algorithms are traced by means of BER for each
Signal to Noise Ratio (SNR = Eb /N0 ). Coding gain comparisons are achieved refer-
encing the case where the transmission is performed without encoding on the trans-
mitter side and without decoding on the receiver side. In this case, Bit Error Rate
achieves 10–5 according to an SNR = 9.6 decibels through AWGN channel. This last
is regarded like binary channel exploiting BPSK modulation.
In Fig. 1(a), (b) and (c), the PHR-SDHT decoding performances are presented
for certain BCH codes with lengths from 31 to 127. These figures show the presence
of a significant coding gain. For instance, the coding gain is superior to 4 dB for
BCH(31, 16, 7), nearly 4 dB for BCH(63, 45, 7) and nearly 3 dB for BCH(127, 113, 5).
In Fig. 2(a) and (b), results comparison between PHR-SDHT, Chase-2 [1]
and Chase-HSDec [15] decoding algorithms is presented for BCH(31, 16, 7) and
BCH(31, 21, 5). From these figures, we deduce that our proposed decoding algorithm
guarantees better performances than the two other decoders do.
Table 1 Experiment
Channel Additive White Gaussian
parameters
Noise
Modulation Binary Phase Shift Keying
(BPSK)
Minimum transferred blocks 1000
Minimum residual errors 200
786 H. Faham et al.
BER
-3
10
-4
10
-5
10
1 2 3 4 5 6 7
SNR
(a)
-1
10
BCH(63,57,3);s=3
BCH(63,51,5);s=5
BCH(63,45,7);s=4
-2
10
BER
-3
10
-4
10
-5
10
1 2 3 4 5 6 7
SNR
(b)
-1
10 BCH(127,120,3);s=2
BCH(127,113,5);s=3
-2
10
BER
-3
10
-4
10
-5
10
1 2 3 4 5 6 7
SNR
(c)
High Performance Decoding by Combination of the Hartmann Rudolph Decoder … 787
In Fig. 3, results comparison between PHR-SDHT, SD1 [20], GADEC [4] and
PHR-ARDecGA [18] decoders is made for BCH(63, 51, 5). The comparison results
confirm that PHR-SDHT exceeds its competitors.
With the aim of showing our serial concatenation effectiveness, a comparison of
the suggested scheme PHR-SDHT and Hartmann Rudolph algorithm is presented by
means of Fig. 4 for BCH(31, 21, 5) code. This comparison indicates that the proposed
combination of Hartmann Rudolph partially exploited and SDHT guarantees great
performances using a small part of the dual codewords. The obtained good results are
based on just 218 dual codewords, which means that we have employed just 21.29%
from dual code space. Table 2 presents the reduction rate of used dual codewords.
Chase-2
between PHR-SDHT, Chase-HSDec
Chase-2 and Chase-HSDec PHR-SDHT;s=7
for: a BCH(31, 16, 7) and 10
-2
b BCH(31, 21, 5)
BER
-3
10
-4
10
-5
10
1 2 3 4 5 6 7
SNR
(a)
-1
10
Chase-2
Chase-HSDec
PHR-SDHT;s=5
-2
10
BER
-3
10
-4
10
-5
10
1 2 3 4 5 6 7
SNR
(b)
788 H. Faham et al.
-1
SD1
10 GADEC
PHR-ARDecGA
PHR-SDHT;s=5
-2
10
BER
-3
10
-4
10
-5
10
1 2 3 4 5 6 7
SNR
Fig. 3 Results comparison between PHR-SDHT, SD1, GADEC and PHR-ARDecGA for
BCH(63, 51, 5) code
-1
10
HR
PHR-SDHT
-2
10
BER
-3
10
-4
10
-5
10
1 1.5 2 2.5 3 3.5 4 4.5 5 5.5 6
SNR
To make sure of the temporal efficiency of our proposed decoder, we have plotted
in the Fig. 5 the ratio of the required run time of PHR-SDHT to the one of HR
algorithm for BCH(31, 16, 7). The reduction in run time is between 80 and 84%. We
deduce that PHR-SDHT has allowed us to minimize in a very large way the run time
of the HR algorithm. This proves the efficiency and the rapidity of our concatenation
scheme.
High Performance Decoding by Combination of the Hartmann Rudolph Decoder … 789
0,18
0,17
0,16
1 1.5 2 2.5 3 3.5 4 4.5 5 5.5 6
SNR
5 Conclusions
References
1. Chase D (1972) A class of algorithms for decoding block codes with channel measurement
information. IEEE Trans Inf Theory 18:170–181
2. Hartmann CRP, Rudolph LD (1976) An optimum symbol-by-symbol decoding rule for linear
codes. IEEE Trans Inf Theory 22:514–517
3. Han YS, Hartmann CRP, Chen CC (1991) Efficient maximum likelihood soft-decision decoding
of linear block codes using algorithm A*, Technical Report SU-CIS-91-42, School of Computer
and Information Science, Syracuse University, Syracuse, NY 13244, December
4. Maini H, Mehrotra K, Mohan C (1994) Soft-decision decoding of linear block codes using
genetic algorithms. In: Proceedings of the IEEE international symposium on information theory,
Trondheim, Norway, p 397
5. Cardoso FA, Arantes DS (1999) Genetic decoding of linear block codes. In: Proceedings
of international conference on telecommunications congress on evolutionary computation,
Washington, DC, USA, pp 2302–2309
6. Azouaoui A, Chana I, Belkasmi M (2012) Efficient information set decoding based on genetic
algorithms. Int J Commun Netw Syst Sci 5(7):423–429
7. Azouaoui A, Belkasmi M, Farchane A (2012) Efficient dual domain decoding of linear block
codes using genetic algorithms. J Electr Comput Eng. Article ID 503834
790 H. Faham et al.
Abstract People around the world often rely on online community forums for
answers to their queries. These forums have become hugely popular in the last decade,
leading to a spurt in the number of users and questions. For a better user experience,
quality monitoring is essential. However, manual moderation of millions of questions
is infeasible. Prior works mostly rely on handcrafted features which is ineffective
or use community feedback as part of learning which makes them unsuitable for
monitoring during question creation. In this work, we use recent deep learning tech-
niques to assess the quality of questions in online community forums at creation time.
We evaluate our model on the StackOverflow dataset that contains 60000 questions
across three qualities. Our model achieves an F1 score of 0.92 on this dataset.
1 Introduction
Online community forums allow for the exchange of knowledge between people
across the world. Over the past decade, an exponential rise in smartphones and
Internet penetration has made these online forums accessible to people across the
world. Question & Answer forums allow people to get answers directed specifically
to their question, rather than having to search haphazardly from a myriad of resources.
Moreover, users are able to get answers from experts in their field at free of cost.
These advantages of online forums have resulted in their substantial growth, with an
enormous number of users across the world relying on these sites for solutions [1]
(Fig. 1).
The rise in users of online forums has naturally led to a tremendous increase in the
number of questions being raised on these forums. To maintain the standard of these
questions, they are either checked manually for their quality through moderators or
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 791
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_72
792 H. Rithish et al.
Fig. 1 In this paper, we are interested in classifying user-generated questions on online forums
automatically, based on its quality
is left to the community to flag questions of poor quality. However, such manual
moderation on millions of questions is ineffective and infeasible. Questions that are
of low quality often result in users spending more time understanding them, either
by asking clarification or by editing the question. Further, questions with low-quality
have been observed to receive answers of poorer quality [2]. In contrast, high-quality
questions receive more constructive answers which also result in long-term value with
users being able to refer them in the future. Overall, low-quality questions reduce the
productivity of users on the site. Thus, we need an automated solution that can flag
potentially low-quality questions. Such a system can help analyze the quality of the
question during creation, and suggest the user to improve the quality before posting
them. They can also be used as input to ranking mechanisms during question search.
Forums such as Quora, Reddit, Stack Exchange, Yahoo Answers and Wiki
Answers serve the general audience and contain discussions on a variety of domains.
Along with such forums, there co-exists another spectrum of online forums that
are dedicated to a particular community, such as Brainly (academic), Avvo (legal),
Sharecare (health), and MadSci Network (Science). Here, we focus on questions
from the StackOverflow website, which is a dedicated forum to discuss queries in
programming. StackOverflow have tried maintaining the quality of these forums by
asking users to follow certain guidelines1 . Despite that, Correa and Sureka [3] esti-
mate around 293 K questions from the period 2009 to 2013 were deleted due to being
off-topic or containing low quality content.
In this work, we automatically assess the quality of StackOverflow questions
through recent deep learning techniques. We work with a dataset of 60000 questions
collected between 2016 and 2020 and released by the StackOverflow platform. We
develop an end-to-end pipeline that pre-processes the input question, extracts relevant
features and classifies them into one of three categories. We focus only on inputs that
an user has access to at creation time, allowing for instant flagging and suggestions.
1 http://stackoverflow.com/help/how-to-ask.
Automated Assessment of Question Quality … 793
In the remaining part of the paper, Sect. 2 discusses prior works on text classi-
fication and question quality assessment, Sect. 3 provides information on architec-
ture, Sect. 4 elaborates on dataset used and implementation details, Sect. 5 shows
experiments conducted and their corresponding results, and Sect. 6 concludes our
work.
2 Related Work
Question classification takes its root in text classification. This domain of text clas-
sification began when IBM stared investigating into techniques for information
retrieval in the 1950’s. One of the earliest works in this domain were by Maron
and Kuhns [4], where they relied on keywords to classify text. In the following
decades, researchers designed classification methods that only worked for a specific
field. These methods were designed with the knowledge of experts in the field by
defining field-specific rules. Naturally, these approaches did not generalize well to
other fields and started losing popularity when machine learning methods gained
prominence in the 1990’s. Machine learning was used to extract features instead of
relying on rules and then classify the text into relevant categories, based on these
features. Some of the earlier machine learning techniques used for text classification
include Naive Bayes [3], Nearest Neighbour [6], Linear Discriminant Analysis [7],
Decision Trees [8], Logistic Regression [8] and Support Vector Machines [9].
Techniques in classical machine learning for text classification evolved over the
next two decades, leading to practical deployment of these systems for a limited
number of use-cases. However, these methods did not scale well to large datasets
or content with complex semantics. With the rapid growth of deep learning in this
past decade, however, text classification systems have matured significantly and
are nowadays being deployed widely. Specifically, various improvements in vanilla
Recurrent Neural Networks have facilitated this progress in text classification [11–
13]. While all the above methods discuss several approaches for text classification,
we specifically focus on the task of assessing questions based on its quality.
Agichtein et al. [14] consider inputs from both the question and their answers. They
establish that there is a strong relationship between the quality of a question raised
and their feedback from users. Shah et al. [15] approach this task differently, by first
categorizing a question and then extract category-specific features for estimating
the quality of a question. Ponzanelli et al. [16] work on identifying good quality
794 H. Rithish et al.
questions that have been mistakenly flagged for review. All the above approaches,
however, rely on user answer information or community feedback for estimating
question quality.
Tausczik and Pennebaker [17] show that a user’s reputation influences the quality
of a question. They design their model using votes received by a question as a feature
to determine question quality. Ravi et al. [18] give importance to the length of the
question content and title. They extend [17] by normalizing votes received by a
question with the topic’s popularity. In both these approaches, they use information
outside of what is available at question creation time, which makes them ineffective
to provide immediate feedback while a user drafts a question. In contrast to all the
above approaches, we use recent techniques in natural language processing along
with inputs (title, body text, tags) that are available at creation time to assess the
quality of a question.
See Fig. 2.
Fig. 2 We start by concatenating multiple components of our question into a single input. Our
combined input is then passed through a set of preprocessing steps. Word embeddings are obtained
for the processed text and then fed to a two-layer Bidirectional LSTM. The extracted features are
then passed to a Fully Connected layer. Finally, we assess the quality of the input question by
passing through a Softmax layer
Automated Assessment of Question Quality … 795
In this section, we define the quality assessment of community questions task. Given
an input title H, body text B and a set of tags T = <t1 , t2 ,… t|T| >, our goal is to classify
the question into one of C qualities.
We start by concatenating multiple components of our question, viz., title, body text
and associated tags into a single input, making it easier for the model to process
further. We then filter out special characters and punctuation present in the input.
We follow this by tokenizing the input, based on the top K most frequent words in
the training set. We finally limit the size of our question to a maximum of L words,
by truncating if the question is longer than L words or by padding if the question is
lesser than L words.
Each word in the question is first converted to its corresponding word embedding.
Embeddings are a representation of text, with each word being encoded as a feature
vector. We employ Global Vectors for Word Representation (GloVe) [19] for this
purpose. GloVe is an unsupervised learning algorithm for obtaining vector repre-
sentations for words [19]. GloVe does not rely just on the local context, but also
incorporates global statistics to obtain word embeddings. The word embeddings are
trained on aggregated global word-word co-occurrence statistics from a corpus. In
this work, we pre-train the model on the Twitter corpus and obtain a 25-d feature
vector for each word.
3.4 Model
to the end and another from the end to the beginning. Through this, we obtain two
hidden states, one with past context and another with future context, which we then
combine through concatenation. We use a hidden state of 128 units and thus, after
concatenation, we obtain a 256-d feature vector. Additionally, we incorporate dropout
and recurrent dropout to minimize overfitting. The sequence of L outputs from the
earlier computation, are passed through another layer of Bi-LSTM to increase the
complexity of the model function. We retain the same structure for the second layer,
and only tweak hyper-parameter values. We pass the output of the 2-layer Bi-LSTM
through a Fully Connected layer, whose output forms the final feature vector for
our input question. The Fully Connected layer transforms the 256-d feature vector
into a |C|-d vector, with each unit representing the unnormalized log value of a its
classification label. Finally, we obtain the classification labels for the question by
passing the unnormalized log values through a Softmax layer.
Since our task is classification, we employ the standard Cross Entropy loss.
4.1 Datasets
Stack Overflow is a widely used question and answering site for programming ques-
tions. The site is a community-based platform, where users can raise new questions
or provide answers to existing questions of other users. Recently, Stack Overflow has
collected a dataset of 60000 questions between 2016 and 2020 and released the data
publicly. A question consists of a title, body and one or more tags. The task involves
classifying a question into one of the three categories:
• HQ: High-quality posts with 30+ score and without a single edit.
• LQ (Close): Low-quality posts that were closed by the community without a single
edit.
• LQ (Open): Low-quality posts with a negative score and with multiple community
edits. However, they still remain open after the edits.
The dataset is balanced, with 60000 questions being evenly distributed between the
three categories. The training and validation set contain 45000 and 15000 questions
respectively, with the questions from each category being sampled in a stratified
manner (Fig. 3).
Automated Assessment of Question Quality … 797
For tokenization, we set the number of words K = 10000. We limit the number of
words in each question, L to 300. For the Bi-LSTM module, we set dropout and
recurrent dropout to 0.4 for the first layer and decrease the value to 0.2 for the second
layer. We use Adam for optimization, with an initial learning rate of 10−2 and reduce
the learning rate by a factor of 0.4 whenever the loss plateaus. We train the model
on an NVIDIA-1080 Ti GPU for 15 epochs, with a batch-size of 256. We show the
training plots of our model in Fig. 4.
798 H. Rithish et al.
Fig. 6 Qualitative examples of HQ questions that are correctly predicted by our model
Fig. 7 Qualitative examples of LQ (Close) questions that are correctly predicted by our model
Fig. 8 Qualitative examples of LQ (Open) questions that are correctly predicted by our model
6 Conclusion
There has been a rapid growth in users and their queries in online community forums,
necessitating an automated quality control mechanism during question creation time.
In this paper, we design a model that takes title, body text and tags associated with a
question as input and predicts the quality of a question. Our model preprocesses the
inputs, embeds them into word vectors using GloVe, extracts features using a 2-layer
deep bi-directional LSTM and finally uses a Softmax layer for classification. We
evaluate our model on the StackOverflow dataset that consists of 60000 questions
800 H. Rithish et al.
and 3 categories of quality. Our model achieves an F1-score of 0.92, indicating that
our system can be deployed for practical use.
References
1 Introduction
Query facets categorize web searches based on some attributes. Faceting technology
facilitates the user to narrow down the web searches based on their interest. It is
widely used in e-commerce websites, online libraries, and social networks. Facets
are generated based on the web content and they describe the properties of the
information. Most recent systems [1] have used query facets in order to achieve
feasible and promising recommendation results in the era of Semantic Web. The
web information for the search undergoes text analysis to identify the latent patterns.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 801
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_73
802 D. Surya et al.
2 Related Work
Jiang et al. [1] used knowledge bases to mine query facets. Their approach covers a
wide range of information to generate query facets rather than only the top results
of the web search. Ramya et al. [2] put forth a method to extract facets for user
queries automatically. The lists are classified to generate the user query facets based
on HTML tags, free-text patterns, and repeat regions. Cosine Similarity is employed
to evaluate the similarity between the lists. The High-Quality Clustering algorithm
classifies the items, and the best items in each cluster are recommended to the user
QFRDBF: Query Facet Recommendation Using Knowledge … 803
as the facets. Vandic et al. [3] put forth a framework dynamic facet ordering in
e-commerce web applications. An automated algorithm ranks facets based on speci-
ficity and dispersion measures. Some specific issues to e-commerce, such as copious
numerical facets, classification of facets by their mutual attributes, and the chance of
multiple clicks, are also addressed by their work. Hervey et al. [4] study the extrac-
tion and ranking of facets in geospatial data portals. Their study proposes methods
of alternative interfaces and query processing pipelines to enhance the search for
geospatial data. Babu et al. [5] present a method to improve the user search experi-
ence by incorporating sub-queries and facets. Galán et al. [6] presented evaluation
approaches to compare region query for the DBSCAN algorithm and a novel region
query strategy. The overall number of region queries and the nearest neighbor search’s
complexity for each region query is minimized. The query strategies in each region
are comparatively assessed based on clustering effectiveness and efficiency. Cherrat
et al. [7] proposed an technique to advance the quality of segmenting the finger-
print images by employing K-means and DBSCAN algorithm. The quality of the
fingerprint images are enhanced by the application of Soble and TopHat filtering algo-
rithms. K-means clustering algorithm is used to obtain the precise segmentation of the
foreground and background region for each block in the image. The computing time
is reduced by using the local variance thresholding technique. Upon implementing K-
means clustering, DBSCAN clustering is applied to overcome the flaws of the former
one. Xue et al. [8] proposed a technique to match biomedical ontologies using the
firefly algorithm. Their work introduced the compact firefly algorithm (CFA), which
outperforms the existing state of art swarm intelligent algorithms. Kaushal et al. [9]
proposed a segmentation technique based on Firefly optimization to segment medical
images of breast cancer irrespective of the image type or modality. Their approach
is compared with the results obtained from the current advanced techniques, and the
results obtained certifies the performance of the proposed method. Xie et al. [10]
addresses the two main problems presented by the K-means clustering model: the
initialization sensitivity problem and local optima traps. These problems are fixed
by the methods proposed by the author in this paper Iwendi et al. [11] proposed
a TF-IDF algorithm with the temporal Louvain method to analyze the text from
various intelligent sensing systems. The proposed work enables analysts to make
reliable decisions by classing documents into hierarchical structures and presenting
the relationships among the variables in the document. The results obtained vali-
date the proposed work’s accuracy and shorter execution time over other existing
approaches. In [12–26] several semantic and ontological approaches in support of
the existing literature are depicted.
804 D. Surya et al.
Queries that represent the user intention are obtained. The collected queries undergo
query preprocessing to generate query words. The pre-processed dataset and query
words are assessed using TF-IDF and reordered based on the frequency. The TF-IDF
is a statistical measure used to determine how much each word in a document is related
to it. It is implemented in information retrieval, keyword extraction, and automated
text analysis. The term frequency and inverse document frequency convey a particular
word’s frequency in one document and a body of documents, respectively. Most of
the machine learning algorithms do not work with text but works with numbers.
The TF-IDF is used to vectorize the text. Text Vectorization process converts a text
document to a vector or array of numbers. It can represent a document as a vector of
numbers to be fed to a machine-learning algorithm. The term frequency and inverse
term frequency are computed as shown in the Figs. 1 and 2, respectively. The value
of TF-IDF is computed by estimating the product of the TF (term frequency) and
IDF (inverse document frequency).
(1)
(2)
(3)
The rearranged items are grouped based on concept similarity, and a knowledge
centric DBSCAN algorithm is employed to generate facets. DBSCAN is an agglom-
erative clustering algorithm which can create automatic categories and determine the
outliers. Two parameters are required to run the algorithm : the radius (maximum) of
a local neighborhood, the number of points (minimum) needed to grow the cluster.
Minimum domain knowledge is adequate to set the hyperparameters. The number
of clusters formed is not determined before training the dataset. It is very efficient
on large data sets as it works excellent with arbitrary clustering, cut noise, and label
outliers. The samples are mutually exclusive and non-exhaustive. Also, the domain
knowledge is fed to the firefly algorithm and concept similarity algorithm to create
facets.
QFRDBF: Query Facet Recommendation Using Knowledge … 805
4 Implementation
The dataset used in this work is the User Q dataset. The User Q dataset contains
90 queries in XML format, and each of those queries contains 100 documents.
The id, title, URL, description, rank, and text of the document, repeat region,
converted HTML list queries are found in each document. The implementation is
performed in Google’s Collaboratory environment in a computer with i7 processor
and 16 GB RAM. Experiments were conducted for News Domains, and Ontologies
were modelled for the Domains of various News. OntoCollab is used for modeling
and visualization of Ontology. The OntoCollab dynamically models ontologies and
Customized Web Crawlers are used to extract metadata for Ontology Modeling.
Hermit Reasoner is used for reasoning of Ontologies. NLTK framework is used to
perform the Natural Language Processing tasks. The concept similarity (CS) has
been employed to identify the semantically similar concepts in different Ontologies
and measure the similarity between concepts in an Ontology. The neighborhood set
806 D. Surya et al.
and the feature set of a concept are used to determine its similarity with another
concept. The most related concepts to a user query can be obtained using the CS
measure. Concept maps of the ontologies can be exploited to extract the concepts.
Mostly, experts design the concept map with the help of the ontology developed by
the domain experts. The concept maps can also be automatically derived from the
ontologies that are used for comparison. An XML file is created for some specific
classes that comprise the representative and characteristic set of some classes in
Ontology. The characteristics and representatives obtained from the previous step
are utilized to calculate the concept similarity measure value. The obtained concept
similarity score is exploited to determine the most related concept to the user query.
QFRDBF: Query Facet Recommendation Using Knowledge … 807
The queries collected are synonymized using WordNet 3.0. WordNet 3.0 is used
to obtain semantically similar terms to the query, such as synonyms, hyponyms, and
meronyms. DBpedia is exploited to obtain real-world knowledge. Jaccard similarity
808 D. Surya et al.
The clusters obtained are evaluated using PRF, wPRF, Purity, R1, F1, F5, and NMI
to determine its quality and efficacy. Normalized Mutual Information (NMI) is an
external measure used to determine the quality of the clusters. Two step clustering
with different numbers of clusters can be compared using NMI values as it is a
normalized score. The NMI score for a cluster CL with respect to a class label CA is
calculated as depicted in Eq. 4. The similarity between two clusters is measured by
Random Index (RI). RI is a statistical measure used widely in clustering to determine
the accuracy. RI value can be calculated, as shown in Eq. 5. TRP, FLP, FLN, TRN
represents the number of True Positives, False Positives, False Negatives, and True
Negatives, respectively. The F-measure estimates the clusters’ accuracy describing
the class labels from the test’s harmonic mean of the precision and recall. The F-
measure of the cluster CL with respect to a class CA is computed as depicted in
Eq. 6.
(4)
(5)
The weighted total of maximum F-measure score for a cluster C gives the overall
F-Measure value. The overall F-measure value can be computed as shown in Eq. 7.
Of all the generated facets, some are useful than the other. Normalized Discounted
Cumulative Gain (nDCG) is employed to identify and extract the most useful facets
from a collection of facets. The nDCG method can be applied to the facet collection
to rank the facets and identify the useful ones. It is applied as depicted in Eq. 8.
Here, DCGk represents the cumulative gain obtained by correct ordering, and iDCGk
depicts the discounted gain of the ith facet. The nDCG measure can be segregated into
two kinds oF-Measure s: first, purity nDCG (fp−nDCG) and second, recall purity
nDCG (rp − nDCG). The former is based on each cluster’s original representation,
while the latter is based on overall facets. The product of DGi with the correctly
assigned item percentage yields the purity of the ith facet. The weights associated with
each item in fp-nDCG and rp-nDCG are illustrated in the Eq. 9 and 10, respectively.
QFRDBF: Query Facet Recommendation Using Knowledge … 809
(6)
(7)
DC G k
n DC G k = (8)
n DC G i
|C L| ∩ f i
wi = (9)
fi
|C L ∩ f i | |C L ∩ f i |
wi = (10)
fi CL
Table 1 compares the results obtained among the proposed approach QFRDBF,
QDMiner, and AEFTM (Automatic Extraction of Facets for user Query in Text
Mining) frameworks on the user Q dataset. The nDCG score of the QFRDBF is
higher than the nDCG score of AEFTM and QDMiner. The DCG score of QFRDBF,
AEFTM, QDMiner are 0.691, 0.738, and 0.938. The PRF and wPRF score of
QFRDBF is much greater than the other baseline models. The PRF score of the
proposed QFRDBF approach is 0.78 while that of QDMiner and AEFTM are 0.392
and 0.436, respectively. The wPRF scores of QDMiner, AEFTM, QFRDBF are
0.392, 0.436, and 0.78. The purity value of QFRDBF is greater than AEFTM and
QDMiner, as the similarity of the items found in clusters obtained from the proposed
approach is higher than the items belonging to the clusters generated from AEFTM
and QDMiner. The purity values of QDMiner, AEFTM, QFRDBF are 0.912, 0.932,
and 0.972. The proposed approach QFRDBF has a better R1 measure value than
AEFTM and QDMiner. The R1 measure value of QFRDBF is 0.967, while the R1
scores of AEFTM and QDMiner are 0.937 and 0.923. The F1 and F5 measure of
FRDBH are 0.941 and 0.882. FRDBH has higher scores than QDMiner and AEFTM
as 0.735, 0.689 are the scores of the former and 0.792, 0.59 are the scores of the
latter. Also, FRDBH has the highest NMI scores among the other baseline methods,
such as AEFTM and QDMiner. The NMI scores of FRDBH, AEFTM, and QDMiner
are 0.984, 0.819, and 0.878.
6 Conclusions
The queries that capture the user’s interest are collected, and query words are obtained
by preprocessing them. The query words are synonymized using WordNet 3.0, and
the real-world knowledge is obtained from DBpedia. The Jaccard similarity among
the query words, synonymized terms, and real-world knowledge are computed. Based
on the results, the semantically similar terms are identified and collected in a modi-
fied database. The DBSCAN clustering algorithm is employed on the modified
database, and the firefly optimization technique is applied to the result obtained
to generate query facets. The results achieved validates the efficacy and the perfor-
mance of the proposed framework, i.e., Facet Recommendation Using Knowledge-
Centric DBSCAN and Firefly (QFRDBF). It can be inferred from the results that the
proposed framework functions better than the other baseline frameworks. QFRDBF
has the highest PRF, wPRF, Purity, R1, F1, F5, and NMI scores among QDMiner
and AEFTM. A F1 and F5 measure of 91.1% and 88.2% are achieved.
References
1. Jiang Z, Dou Z, Wen JR (2016) Generating query facets using knowledge bases. IEEE Trans
Knowl Data Eng 29(2):315–329
2. Ramya RS, Raju N, Sejal N, Venugopal KR, Iyengar SS, Patnaik LM (2019) Automatic
extraction of facets for user queries [AEFUQ]. In: 2019 Fifteenth International Conference
on Information Processing (ICINPRO), pp 1–6. IEEE
3. Vandic D, Aanen S, Frasincar F, Kaymak U (2017) Dynamic facet ordering for faceted product
search engines. IEEE Trans Knowl Data Eng 29(5):1004–1016
4. Hervey T, Lafia S, Kuhn W (2020) Search facets and ranking in geospatial dataset search.
In: 11th International Conference on Geographic Information Science (GIScience 2021)-Part
I. Schloss Dagstuhl-Leibniz-Zentrum für Informatik
5. Babu D: Web search using automatically generated facets (2019). In: 2019 2nd International
Conference on Intelligent Computing, Instrumentation and Control Technologies (ICICICT),
vol 1, pp 37–40
6. Galán SF (2019) Comparative evaluation of region query strategies for DBSCAN clustering.
Inf Sci 502:76–90
7. Cherrat EM, Alaoui R, Bouzahir H (2019) Improving of fingerprint segmentation images based
on K-means and DBSCAN clustering. Int J Electric Comput Eng 2088–8708:9
8. Xue X (2020) A compact firefly algorithm for matching biomedical ontologies. Knowledge
and Information Systems, pp 1–17
9. Kaushal C, Kaushal K, Singla A (2020) Firefly optimization-based segmentation technique to
analyse medical images of breast cancer. Int. J. Comput. Math. 1–16
10. Xie H, Zhang L, Lim CP, Yu Y, Liu C, Liu H, Walters J (2019) Improving K-means clustering
with enhanced Firefly Algorithms. Appl Soft Comput 84:
11. Iwendi C, Ponnan S, Munirathinam R, Srinivasan K, Chang CY (2019) An efficient and
unique TF/IDF algorithmic model-based data analysis for handling applications with big data
streaming. Electronics 8(11):1331
12. Deepak G, Santhanavijayan A (2020) OntoBestFit: A Best-Fit Occurrence Estimation strategy
for RDF driven faceted semantic search. Computer Communications, vol 160, 284–298
QFRDBF: Query Facet Recommendation Using Knowledge … 811
13. Pushpa CN, Deepak G, Kumar A, Thriveni J, Venugopal KR (2020) OntoDisco: improving web
service discovery by hybridization of ontology focused concept clustering and interface seman-
tics. In: 2020 IEEE International Conference on Electronics, Computing and Communication
Technologies (CONECCT), pp 1–5. IEEE
14. Kumar A, Deepak G, Santhanavijayan A (2020) HeTOnto: a novel approach for conceptualiza-
tion, modeling, visualization, and formalization of domain centric ontologies for heat transfer.
In: 2020 IEEE International Conference on Electronics, Computing and Communication
Technologies (CONECCT), pp 1–6. IEEE
15. Deepak G, Kasaraneni D (2019) OntoCommerce: an ontology focused semantic framework
for personalised product recommendation for user targeted e-commerce. Int J Comput Aided
Eng Technol 11(4–5):449–466
16. Gulzar ZA, Anny L, Gerard D: Pcrs: personalized course recommender system based on hybrid
approach. Procedia Comput Sci 125, 518–524 (2018)
17. Deepak G, Teja V, Santhanavijayan A (2020) A novel firefly driven scheme for resume parsing
and matching based on entity linking paradigm. J Discrete Math Sci Cryptography 23(1):157–
165
18. Haribabu S, Sai Kumar PS, Padhy S, Deepak G, Santhanavijayan A, Kumar N (2019) A
novel approach for ontology focused inter- domain personalized search based on semantic set
expansion. In: 2019 Fifteenth International Conference on Information Processing (ICINPRO),
Bengaluru, India, 2019, pp 1–5. https://doi.org/10.1109/icinpro47689.2019.9092155
19. Deepak G, Naresh Kumar G, Sai Yashaswea Bharadwaj VSN, Santhanavijayan A (2019)
OntoQuest: an ontological strategy for automatic question generation for e-assessment using
static and dynamic knowledge. In: 2019 Fifteenth International Conference on Information
Processing (ICINPRO), pp 1–6. IEEE
20. Santhanavijayan A, Naresh Kumar D, Deepak G (2010) A semantic-aware strategy for auto-
matic speech recognition incorporating deep learning models. In: Intelligent System Design,
pp 247–254. Springer, Singapore
21. Deepak G, et al. (2019) Design and evaluation of conceptual ontologies for electrochemistry as a
domain. In: 2019 IEEE International WIE Conference on Electrical and Computer Engineering
(WIECON-ECE), IEEE
22. Deepak G, Priyadarshini JS (2018) Personalized and enhanced hybridized semantic algorithm
for web image retrieval incorporating ontology classification, strategic query expansion, and
content-based analysis. Comput Electric Eng 72:14–25
23. Kaushik IS, Deepak G, Santhanavijayan A (2020) QuantQueryEXP: a novel strategic approach
for query expansion based on quantum computing principles. J Discrete Math Sci Cryptography
23(2):573–584
24. Varghese L, Deepak G, Santhanavijayan A: an IoT analytics approach for weather forecasting
using Raspberry Pi 3 Model B+. In: 2019 Fifteenth International Conference on Information
Processing (ICINPRO), pp 1–5. IEEE (2019)
25. Shreyas K, Deepak G, Santhanavijayan A (2020) GenMOnto: A strategic domain ontology
modelling approach for conceptualisation and evaluation of collective knowledge for mapping
genomes. J Stat Manage Syst 23(2):445–452
26. Deepak P, Zakir M (2016) Enhanced neighborhood normalized pointwise mutual information
algorithm for constraint aware data clustering. ICTACT J Soft Comput 6(4)
SentiFusion (SF): Sentiment Analysis
of Twitter Using Fusion Techniques
Abstract The Internet has become a domain for learning, idea exchange, feedback,
sharing thoughts or opinions, and many more. In the current world, Twitter has
been the most liable platform where people express their views and ideas about a
various range of products. In this paper, different feature extraction methods are
used to get more accurate sentiment analysis results. Firstly, tweet pre-processing
is performed using Euclidean Vector Length Similarity and other dictionaries. The
feature extraction is performed using (Jaccard Similarity and Trigrams) and (K-
means and KL-divergence). Then using tf-idf and Shannon Entropy, feature fusion
and weighting are done. At last, before final Sentiment Computation, Bagging is
accomplished using SVM classifiers and Decision Trees.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 813
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_74
814 S. Anand et al.
Extraction. It has more than 150 users with 2,400 tweets per day. Analyzing Twitter
is the best option because it contains various opinions, feedback, and reviews.
Extracting fascinating information and non-trivial patterns from unstructured data
is the primary concern of Opinion Mining and Sentiment analysis which are the
branches of Texting mining. Opinion mining (analyzing people’s opinions) is the
automated process of identifying and extracting subjective information from a given
document. The main aim of doing opinion mining is to classify the tweets as per
their sentiment values. The sentiment values will help in making the decision-making
process for different businesses and other applications.
In SentiFusion, we propose a combination of several methods, namely the Jaccard
Similarity, Trigrams, K-means Clustering with Silhouette coefficient, and Kullback
Leibler Divergence, which help in extracting the right features and makes the data
well-structured for classification. Following the feature extraction, the tf-idf has been
encompassed to act as a feature weighting scheme. Bagging is done using the two
most vital algorithms, which are Support-vector classifiers and Decision Trees. The
experiments have been conducted on Sentiment140 dataset models and have been
baselined with other models to show the proposed SentiFusion approach outperforms
different approaches. SentiFusion has achieved better performance measures in terms
of Precision, Recall, Accuracy, and F-Score.
In Sect. 2, Related Work has been described. Section 3 presents the proposed
system Sect. 4 consists of a description of Sentiment Computation. Implementation
with experimental results and Performance evaluation is discussed in Sect. 5. The
paper is concluded in Sect. 6.
2 Related Work
In [1], (Savitha et al. 2018) have proposed measures to extract sentiment variation’s
reasons on Twitter. The author used two approaches, Cosine Similarity and n-gram
similarity with Latent Semantic Analysis, to notice the similarities between tweets
and then concluded the reasons for sentiment variation by getting the highest simi-
larity value. This approach leads to an overall increase in the time complexity of the
system, which can be reduced. In [2], (Amit and Deshmukh 2013) used Pointwise
Mutual Information (PMI) and Chi-Square (X2) for feature selection. Further, for
sentiment classification, Naive Bayes (NB), Support Vector Machine (SVM), and
Maximum Entropy (ME) were used. The approach was mainly Lexicon-based. This
system lagged in the overall accuracy of judgment. POS was introduced as specific
polarity features in [3] by (Apoorv et al. 2011). They used tree kernels for feature
extraction. This model had outperformed the general state-of-the-art baseline but had
lower F-Scores. (Öztürk and Ayvaz 2018), in [4] performed comparative sentiment
analysis of the tweets retrieved and analyzed tweets in Turkish Language. For this
purpose, they used the Turkish sentiment analysis lexicon then simply classified data
using term frequencies. It had lower F-scores even if only the analysis of Tweets is
concerned. The effects of pre-processing were ignored in [5]. In it, the main focus was
SentiFusion (SF): Sentiment Analysis of Twitter … 815
on classification. (Jianqiang and Xiaolin 2017) tried to depict the six pre-processing
methods’ performances while using two feature models and four classifiers on five
different Twitter datasets. In this paper, the accuracy and F1-measure proposed
methods using classification classifiers were improved by expanding acronyms and
replacing negation. It doesn’t change the URLs, numbers, or stop words, which was
why it increased the time complexity of the system. The authors showed that Logistic
Regression and SVM classifiers are less sensitive than Naive Bayes and Random
Forest classifiers when various pre-processing methods were applied. Textual infor-
mation of Twitter messages and sentiment diffusion patterns were fused to perform
better [6]. But apart from the classifications, other sections of the system decreased
the overall F-scores. (Wang et al. 2020) used sentiment reversal phenomenon. They
proposed an iterative algorithm, which was named SentiDiff, to predict polarities of
the sentiment of Tweets. The PR-AUC improvement (Precision-Recall Area Under
Curve) between 5.09 and 8.38% was obtained using this model on Twitter senti-
ment classification tasks. Common sense knowledge (Domain Specific Ontology),
a very innovative sentiment analysis method, was used in [7] by (Ramanathan and
Meyyappan 2019). They created their Ontology based on ConceptNet. The combined
sentiment lexicon approach finally analyzed sentiments. In the final step, concerning
the domains, semantic orientations of domain-specific features were incorporated.
The only problem with this system was it led to an increase in the time complexity
because of the new Ontology. In [10–21] several ontological supporting strategies
have been discussed.
libraries for text mining. For reference purposes, the proposed system architecture
will be called SentiFusion or SF in further mentions.
The gathered data set is fed to the next stage of pre-processing the data. In Tokeniza-
tion, tweets are broken down into words that are termed as ‘tokens’. Tokens are
normalized to standard form (all lower cases) using Euclidean Vector Length Simi-
larity. Acronyms are converted to their full forms using the Normalization process.
Using the stop word dictionary, each word is tagged as ST (Stop-words) or NST
(Non-Stop-words) for their Parts-of-Speech. Using the emoticon dictionary, emoti-
cons are removed and taken in as input to their sentiment values for each tweet. The
URLs are removed from the tweets. The inflected forms of words are grouped to
find out the base term using the procedure of Lemmatization. Each token is now
classified as (word, POS tag, English-word, Stop-word, Sentiment values). Directly,
these tokens with their classifications are fed as inputs to the next step of the process.
Dataset Description: In SentiFusion, the Sentiment140 data set is being used. This
data set has already been extracted using Twitter API after selecting the default
language as “English”. It contains 1,600,000 tweets. The six fields used from the
dataset here in the proposed model are the target, ids, date, flag, user, and text. The
SentiFusion (SF): Sentiment Analysis of Twitter … 817
In this step for Extracting Features, two methods are adopted. In the First Method,
Jaccard Similarity and Trigrams are being used. In the Second Method, K-means
Clustering with the Silhouette coefficient and Kullback Leibler divergence is used.
Tf-Idf is common to both the methods for Classification. Data from the previous
step is fed to Method 1, where the initial level of feature extraction is done. The
data is provided to Method 2, which refines the feature. This step aims to extract and
summarize as many features from the tweets as possible, which will be fused and
classified in the next steps. Using two methods of feature extraction ensures better
accuracy.
Jaccard Similarity
Jaccard Similarity is the method of grouping the same data from two different data
sets. This algorithm is used to increase the accuracy of the proposed system and
shorten the given data sets. The pre-processed data from the previous step is compared
with different tweets and assigned different weights. The whole file is read for various
tweets to make comparisons between them for similar words. Every tweet will be
assigned a Jaccard Similarity ratio given by dividing the size of the intersection of
the two sets by the size of its union. Since we have already normalized our tweet
in the first step, repeating a word in a single tweet is avoided without changing the
results, a distinct feature of SentiFusion. So, by now, for a detailed analysis, all the
tokenized words of tweets will be grouped in arrays with an ID number representing
the number of appearances of the word present in the whole document. Each array
is termed a cluster [8].
Trigrams
An n-gram is collected from a text corpus, a contiguous set of n items from a given
sample of text. Trigrams are a type of n-grams with n = 3. N-grams are used in
text analysis for grouping up different letters of a word. For example, for the string
“abcdefg” trigrams (n = 3) will be efg, def, cde, and so on. These language models
are used in Natural Language Processing Models to reduce the algorithm’s overall
time complexity. In SentiFusion, trigrams are used to group different elements of the
arrays obtained from the previous step. Each array element will be converted into
chunks of 3 and assigned a weight as per the array’s ID number. Then all the chunks
with their weights will be inserted into a stack. This step is done so that there will be
the least number of comparisons in the next step.
818 S. Anand et al.
Kullback–Leibler Divergence
Kullback-Leibler divergence (relative entropy), in short, called KL divergence,
measures the difference between any two probability distributions. Given two discrete
probability distributions, p(x) and q(x), D(p||q), the relative entropy between p and
q is defined in Eq. (1).
From Eq. (1) it can be conclude that, D(p||q) is always non-negative and if p = q
then D(p||q) = 0. It shows the variation of p(x) with respect to q(x). If KL divergence
is small, there is a high similarity between the distributions of two variables, and
the converse of it is also True. This probability distribution is used to assign more
precise values to the classified clusters obtained from the previous step. Every token
of the clusters is now assigned a decimal value (KL—scores) of their related entropy,
which gives a more precise analysis of Sentiments.
TF-IDF
We implement Term frequency-inverse document frequency, a numerical statistic
for keyword extraction. TF-IDF is related to a bag of words (Bagging). The main
idea behind using TF-IDF as a numeric feature generator is that—more importance
should be given to the word that appears more in one text document and less in
another, which eventually becomes useful for classification. In SentiFusion, this
model is used for emphasizing the tokens, which are frequently repeated. The tokens
in SentiFusion are terms, and similarly, each cluster is considered as a document
here. For every token and cluster, a tf value and idf value are calculated, respectively.
When these values are multiplied together, we get a score that is highest for tokens
that frequently appear in the clusters and low for tokens that occur less often in every
cluster, allowing us to find tokens necessary in the cluster. Hence, if the token is
very common, then the value will approach 0; else, it nears 1. Now, the “tf-idf score”
is assigned to all the tokens, and a distinction is made between normal tokens and
keywords for all the clusters. This tf-idf score is multiplied with the KL-scores from
the previous step. This clustered data is now fed to machine learning algorithms for
further feature fusion and classification.
SentiFusion (SF): Sentiment Analysis of Twitter … 819
This step is for feature fusion by combining the scores of all the tokens and keywords
obtained in the previous step of Feature Extraction. The sentiment scores are multi-
plied with the scores obtained from the earlier steps. A final decimal sentiment score
is obtained by the end of this step. The keywords will naturally have higher scores
as they already have a higher combined tf-idf and KL scores. Let these scores of all
the tokens be called “Senti-scores” for future references. All the clusters are fed to
the next step of feature weighting and bagging to summarize the results of sentiment
analysis of the queried topic.
Shannon and Weaver proposed this feature weighting method in 1963. It acts as a
criterion for measuring the degree of uncertainty represented by a discrete probability
distribution. Entropy here measures existent contrasts between different sentiment
datasets. Let p(x) or pij be the probability mass function of the random variable x or
xij .
Steps for feature weighting through Shannon Entropy are:
1. Normalizing the arrays of performance indices to obtain the outcome pij based
on Eq. (2)
xi j
p(x) = pi j = m (2)
i=1 x i j
Where k = ln(m)
1
(1 − E j )
w j = n (4)
j=1 (1 − E j )
3.5 Bagging
Support Vector Machines or SVMs are supervised learning models. For the clas-
sification of data and then performing Regression analysis as per the need, SVMs
analyze the data and recognize the patterns. In this step, using the SVM classifiers,
we are creating a supervised learning model of three classes—strong sentiments,
neutral sentiments, and Weak sentiments. This classification is done for each token
based on Senti-sores (product of tf idf and KL divergence scores). This model is now
trained to accept and classify any new tokens based on Senti-scores. For each of the
three classes, the tokens are placed and interlinked in space in the form of Decision
trees.
SentiWordNet [9], derived from the WordNet database, is a lexical resource used
for opinion mining. It assigns to each synset (a word or a term which are inter-
changeable synonyms) three sentiment scores: positivity (+1), negativity (−1), and
objectivity(0). In this final step after feature selection, a ranking of the selected tokens
is done using the SentiWordNet dictionary. The selected tokens include both positive
and negative rankings. These score values of the products are inputted to SVC and
Decision Tree classifiers in the three classes. The polarity of the word signifies the
characteristic of a given tweet.
SentiFusion (SF): Sentiment Analysis of Twitter … 821
The whole program was run on a system with an i7 Processor/16 GB RAM. It was
connected with the DCX1 server of supercomputing to supercomputers. Stanford’s
Python Natural Language Toolkit has been used for carrying out NLP as well as Pre-
processing Tasks. Pre-processing of blank space and special character tokenizers
were customized. Similarly, a customized algorithm was used for stemming or
lemmatization and stop-word removals.
The performance of the proposed SentiFusion has been computed using Precision,
Recall, Accuracy, and F-Score as the desired matrix. In Table 1, a comparison between
CSNM paper and SentiFusion paper is carried out. The standard formulae of the
respective terms are used to calculate the performance measure.
Table1 depicts a comparison between the existing method-CNSM and the
proposed method—SentiFusion. In total, 2500 data are compared, of which the 500
datasets of CNSM yield 86.87% of Precision, 88.02% Recall, 87.445% Accuracy,
and an F-score of nearly 87.44%. In contrast, we can see a Precision of 94.14%,
97.18% of Recall, 95.66%, and 95.64% of Accuracy and F-Score, respectively in
SentiFusion. We can observe a considerable difference rate in Accuracy and F-Score
of both the models. With the more reliable F-Score, we can conclude that the proposed
system has higher performance measures.
The SentiFusion method performed better than the CSNM method. This happened
mainly because of the Bi-way verification system in SentiFusion. In the CNSM
method, the author employs n-gram similarity matching and Cosine similarity using
Latent Semantic Analysis methods. This made the model a bit slower in terms of
Time complexity. Topic modeling is used with ranking for feature selection. There
is no robust classifier included in the existing system of CNSM. The proposed
SentiFusion has a concrete Bi-Classifier and Ensemble Classifier with SVM that
harvests boundary features, granular and sub-granular features, unlike LSA. Whereas
in feature extraction of the SentiFusion model, two different methods are used, which
give a higher accuracy rate. The time complexity of the SentiFusion model is also
less as the comparisons are less because of Shannon Entropy.
6 Conclusions
A SentiFusion model was trained which can now be used to analyze sentiments
for any topic properly. The model was trained using two different methods. The
first method consisted of Jaccard Similarity and Trigrams while the second method
had K-means Clustering and KL divergence implemented in it. Then feature fusion
and weighting were facilitated by tf-idf and Shannon Entropy. Bagging in the last
step was done using SVM Classifiers. This model was found to give an accuracy
of 92% in extensive datasets whereas 95% accuracy was easily obtained for smaller
datasets. The average F-score of our model is 94.23 which is higher than the others.
The proposed model has 95% F-score for smaller datasets and 92% for extensive
datasets. This model also has very high precision value. Hence, it is concluded that a
two-way Feature Extraction combining two different methods provides better results.
References
1. Shirbhate AG, Deshmuk SN (2013) Feature extraction for sentiment classification on twitter
data. Int J Sci Res (IJSR), pp 2184–2185
2. Mathapati S, Anil D, Tanuja R, Manjula SH, Venugopal KR (2018) CNSM: cosine and n-gram
similarity measure to extract reasons for sentiment variation on Twitter. Int J Comput Eng
Technol (IJCET) 9:150–161
3. Agarwal A, Xie B, Vovsha I, Rambow O, Passonneau R (2011) Sentiment analysis of Twitter
data. In: Proceedings of the workshop on languages in social media, LSm 2011, pp 30–38
4. Öztürk N, Ayvaz S (2018) Sentiment analysis on Twitter: a text mining approach to the Syrian
refugee crisis. Telemat. Inform. 35(1), 136–147
5. Jianqiang Z, Xiaolin G (2017) Comparison research on text pre-processing methods on Twitter
sentiment analysis. IEEE Access 5:2870–2879
6. Wang L, Niu J, Yu S (2020) SentiDiff: combining textual information and sentiment diffusion
patterns for Twitter sentiment analysis. IEEE Trans Knowl Data Eng 32(10):2026–2039
7. Ramanathan V, Meyyappan T (2019) Twitter text mining for sentiment analysis on people’s
feedback about Oman tourism. In: 2019 4th MEC international conference on big data and
smart city (ICBDSC), Muscat, Oman, pp 1–5
8. https://python.gotrained.com/nltk-edit-distance-jaccard-distance/#Example_1_Character_
Level
9. https://github.com/aesuli/SentiWordNet
10. Deepak G, Santhanavijayan A (2020) OntoBestFit: a best-fit occurrence estimation strategy
for RDF driven faceted semantic search. Comput Commun 160:284–298 (2020)
11. Pushpa CN, Deepak G, Kumar A, Thriveni J, Venugopal KR (2020) OntoDisco: improving web
service discovery by hybridization of ontology focused concept clustering and interface seman-
tics. In 2020 IEEE international conference on electronics, computing and communication
technologies, pp 1–5, July 2020
SentiFusion (SF): Sentiment Analysis of Twitter … 823
Abstract Nowadays, the automotive sector needs more and more digitalized solu-
tions, and this is especially with regard to PLM software and its effectiveness in
communicating information related to new products under development, even if this
is the main role of PLM but to increase its effectiveness and efficiency of collaboration
between automotive companies, their subcontractors and suppliers. Because effective
collaboration leads directly to the rapid success of product lifecycle management.
In this paper we will show the need to better digitalize the communication between
the different actors of the PLM wheel and we will also mention three principles that
will help us to better collaborate in the context of product lifecycle management.
1 Introduction
In a world where the only constant thing is change, the platform Product Lifecycle
Management PLM has been created and designed over the past two decades to
manage the entire product lifecycle efficiently and cost-effectively from ideation,
design, manufacture and then service to disposal. In fact, the huge need of an inno-
vative solution as PLM systems comes from the rapid evolution of the production
worldwide. One reason that pushes organization to adapt their strategies to the context
and the new challenges such as the international concurrence, the segmented market,
the big flow of information and data, the intensive resources difficult to manage and
control, etc.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 825
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_75
826 N. Tilioua et al.
2 Literature Review
It is thanks to the rapid and advanced evolution of information technology that the
automotive sector has become more competitive in recent years. Internet technology
has contributed to better linking and virtually connecting processes even with distant
actors and to the development of new strategies for the development of products
that meet customer requirements and aim to better satisfy them. And as Mr. Sharma
lied in his article, the real challenge of technology lies in its ability to link people,
processes and information [1].
The current state of PLM software is based mainly on the BOL and MOL phases,
but according to existing research, the future vision of PLM will also give importance
to the EOL phase.
The literature sets up several challenges and opportunities of PLM, including
data and information management issues [2] and requirements management [3].
The definition of PLM presented by Corallo [4] groups together the information
and collaboration management of products throughout their life cycle as well as the
inherent aspects of managing. However, it should be noted that providing and sharing
information will enable effective PLM processes, both within an organization and
with partners and stakeholders across the enterprise.
The evolution of PLM also addresses the increasing complexity of companies,
which is mainly due to the following factors increasing organizational growth
(collaboration, decentralization) increasingly complex products [5].
Product lifecycle management systems play an important role in the integration
and process-based organization of information about a series of development and
evolution phases of products, processes and resources.
New information technologies, such as Digital Twin, are encouraging industrial
digital transformation and effective lifecycle management that must now take into
account all phases of the project, where The use of the 3D model in some of them is
no longer the best solution [6].
Digital Twin (DT) has a great capacity to realize Industry 4.0. to simulate working
conditions in real time and make intelligent decisions, when a cost-effective solution
can be easily provided. Knowing that the term PLM allows an application of the
Digital Twin as companies are moving towards green manufacturing with the Digital
Twin which enablesmaximizing production quality throughout a complete loop cycle
[7]. Thanks to the simulations of the Digital Twin tool and the opportunity it gives to
the actors of the value chain to make decisions relevant to the topics concerning the
life cycle of products, based on this opportunity, they canmake intelligent products
with self-awareness [8].
828 N. Tilioua et al.
The Automotive products lifecycle begins with the Research and Development stage.
At the first stage, it is necessary for the vehicle company’s engineers and technicians
to search for the composition of the products,and we know that suppliers develop
new references of these products after any EOM negative return or any innovation of
that product, so the idea is to link and facilitate the references searching to OEM engi-
neers and technicians commands directly with suppliers by using PLM tool connected
with their personal computers, as well as the future manufacturing technology. At
this stage, applications are filed within patent offices to protect the intellectual prop-
erty created during the drug development. It will also require the organization of
pilot production and procurement since at the development stages, sample batches
of automotive products are always required in those tests. It is also necessary to
conduct a search and optimal supplier name selection for the product being developed
with the subsequent necessary measures implementation for registering a command.
Procurement stage is followed by banc tests of parts required.
Once the idea research stage of a new part project is completed, a feedback is made
to OEMs and suppliers. At this stage of the product life cycle, a future study protocol
is created, and after its approval, research and development studies are conducted
in accordance with the requirements of good practice. Component data of product
samples from the research and development stage.
This stage is followed by the registration of the product in the PLM, during which
the actors are responsible for their modifications. After the evaluation, the engineers
issue a marketing authorization for the automotive product. The process of collecting
the documents in the file is complicated and requires specialists with a high technical
level. At this stage, it is necessary to ensure that all the data generated by the previous
steps are integrated.
The production phase normally takes place after registration, but its prepara-
tion takes a considerable amount of time. The organization of the production is
usually done in parallel with the previous steps. The production of automotive parts
is controlled by testing, and the main component of this system is quality assurance
at all stages of production, from the purchase of raw materials and materials and their
control through to the final product.
To remain competitive in today’s digital age, knowledge sharing and knowledge
reuse between different companies in internal or external collaboration needs to be
strengthened, and the knowledge needs of the different phases of the product lifecycle
need to be identified and detailed in a clear way.
We will compare three articles in this sense, the first article chose to deal with the
two BOL phases of the R&D companies because this phase is well known as the most
important stage on which the entire product life cycle is based, and the MOL phase
with logistics companies and it was noticed that the staff in both phases insist on the
need for more advanced integration of the suppliers this is shown in the article by the
following sentence: The R&D staff focused on new materials and innovations from
Using Internet of Things … 829
suppliers to apply them more quickly to their own product development (M2), while
the logistics staff focused on the transport capacity of suppliers because sometimes
their own fleet cannot meet the customers’ requirements.
And for the same article the author focused his vision on expertise as a type of
knowledge to be shared to better streamline and help manage product lifecycles,
saying that R&D expertise was mainly related to design, development, technology
and manufacturing process. For the logistics people, it was mainly related to import
and export, and export, insurance, and political and legal aspects. This by insisting
in the side of logistics companies that the expertise acquired from competitors is
more useful because it is more relevant to them unlike R&D companies which have
confidentiality in their expertise compared to their competitors. And the latter can
only obtain knowledge through other channels, especially through person-to-person
communication. However, logistics personnel expect the government to organize
special meetings to obtain faster updates and among the things that show the need
for supplier integration in logistics are hygiene, safety and temperature in transport,
limiting the weight and length of aircraft.
And an example was given in this article that knowledge sharing is done in a hier-
archical way from the internal base of the company to the external side by specifying
reference persons and insisting that Knowledge must be shared within the depart-
ment, within the company, with the branch, with the supplier, with the customer and
even with the competitor.
In the other article called: A Green PLM approach, which focuses on product
lifecycle management from an environmental perspective, his contribution suggests
a framework for sustainable product development that takes into account the entire
product lifecycle and will enable companies to be more resource efficient.
For him the Factory of the Future is a set of competent regional small and medium
sized companies that could compete but, at the same time, be available to collaborate
when needed by sharing knowledge and resources. with customers, stakeholders and
suppliers and thus enable sustainability through green products and processes (Green
Factory certified).
The author in this article emphasizes a special attention to green manufacturing
which must be considered critical as it is the most polluting stage of the product
life cycle due to greenhouse gas emissions and liquid and solid waste that are not
sufficiently controlled.
In the 3rd article: Convergence of IoT and product lifecycle management in
medical health care, the author highlights the technological and artificial side to
better develop the role of PLM while saying that with the modern cloud and IOT
Intelligent connected devices improve collaboration and flexibility throughout the
value chain, Greg Cline [9] presents that things already generate more data than
people or applications.
The main contribution of this paper is to raise the importance of IoMT integra-
tion with Product Lifecycle Management (PLM), information and data sharing, and
collaboration from medical experts to patients and between devices.
830 N. Tilioua et al.
The author thus insists on the link between the Internet of Things (IoT) and PLM
in the medical market because with Information Communication Technologies (ICT)
and sensors that are driving a dynamic change in industry and academia. As research
on PLM and IoT has been conducted for a long time. But these feature-based devices
consume more energy and a majority of small nodes are battery-powered, which is
a critical and interesting challenge.
The e-Cloud with IoT-enabled devices is being developed; in which devices can
communicate with each other while sharing and exchanging very important informa-
tion and the power and life cycle of these devices will be managed from the beginning
to the end of the medical treatment process thus keeping in mind the ability to gather
knowledge and information.
Finally, I think it is important to combine these three ideas of IOT and lean
manufacturing in green manufacturing by ensuring supplier integration and effective
collaboration between all the collaborators in the value chain with the control of
environmental experts and industrial managers through the use of intelligent sensors
and connecting them directly to the devices of all these people each on their own
side of the collaboration.
Giving the importance to collaboration in the purpose of having the effective
management of products throughout their life cycle. The literature on the efficiency
of integration and communication between suppliers and stakeholders and OEMs its
objective in this article is facilitating the selection of the latest references that can be
adapted to the context of product life cycle management. In Table 1, we summarize
their relevance and importance for decision support in the context of LCM.
And for the problem of efficient supplier integration and communication between
test engineers on new vehicle projects and suppliers of the latest 3D part modification
or creation, I propose to integrate one more step in the PLM tool process which is the
effective communication to order parts directly without the need to switch to another
system to order them, and these steps are outlined as follows (Fig. 1):
As a result of this proposition there will be an efficiency in rapid exchange of
references flows between suppliers and OEMs, and also in logistics. to approve that
we.
The hyper-connected world of products, human actors, and operating environ-
ments, enabled by IoT technologies [13], creates a potential for explosive growth
in the generation of product—related data streams. It is not sufficient anymore to
seek to integrate such information centrally. Instead, part of the information inte-
gration is best performed at the point of data consumption. For such integration to
produce meaningful results, the complexity of relevant product information needs
to be managed. Directly relevant to establishing efficient architectures, indexing,
and big data management capabilities for IoT—generated data is the key concept of
context, adopted in context aware computing [13].
Table 1 Collaboration support issues and principles
Issues Example Principles
Current PLM systems must enable collaboration, C.Vila, An approach to Sustainable Product See Principle 1 below Wherever possible,relevant
Using Internet of Things …
mandatory consultations, and capture inputs, Lifecycle Management (Green PLM), Procedia communications should be captured at an
contributions, and changes in all system Engineering 132 (2015) 585–592 [10] appropriate level of detail;
interactions it can be used to enable better knowledge retrieval.
Collaboration 4.0 concepts can provide a context
for knowledge reuse throughout the product
lifecycle, and time and money optimization
A number of participants can contribute to PLM YanXin, Dealing with Knowledge Management See Principle 2 below
decisions—groups, deparments, and otherrelevant Practices in Different Product Lifecycle Phases In the stages of an automotive part Development
stakeholders within the organizational structures within Product-service Systems, Volume 83, 2019, process, involve the relevant participants in these
and hierarchies of different Pages 111–117 [11] stages and give each participant his or her role and
the opportunity to send messages and order emails
directly to the suppliers of these parts. Assign a
participant as a key decision-maker
Problems using PLM manifest themselves in ALIHASSANSODHRO, Convergence of IoT and See Principle 3 below To improve the integration of
many forms different product lifecycle management in medical actors and minimize training expenses, and also to
healthcare, 2018 [12] ensure the integration of suppliers who are a
different organization by adding a training page in
the PLM tool according to the user’s role
831
832 N. Tilioua et al.
4 Conclusion
This work contributes to provide the collaboration in PLM. Time optimization and all
actors’ integration in automotive company’s value chain are essential conditions to
have an efficient product management. The paper highlights how supporting concepts
in-group collaboration and integration facilities can help to meet the requirements of
choice and reuse of PLM decisions and decision-making processes, thus contributing
to organizational knowledge accumulation and organizational learning. The princi-
ples presented to support decision-making in a PLM context provide a foundation for
effective PLM decision support with significant potential for integration with existing
PLM systems and for exploiting emerging and future PLM ontologies, which may
have broad implications for organizations’ PLM processes and protocols in thefuture.
References
1. Sharma (2005) Collaborative product innovation: integrating elements of CPI via PLM
framework 37(13), 1425–1434
2. Marra M (2012) Supply chain knowledge management: a literature review 39(5): 6103–6110
3. Silventoinen A (2014) Challenges of information reuse in customer-oriented engineering
networks 34(6):720–73
Using Internet of Things … 833
4. Corallo A (2013) Defining product lifecycle management: a journey across features. Def
Concepts, Volume 2013 |Article ID 170812
5. Meier U (2017) Twenty Years of PLM–the good, the bad and the ugly. Product Lifecycle
Manage Ind. Fut, pp 69–77
6. Tchana Y (2019) Designing a unique Digital Twin for linear infrastructures lifecycle
management. Procedia CIRP84, pp 545–549
7. Papinniemi J, Hannola L, Maletz M (2014) Challenges in integrating requirements management
with PLM. Int J Prod Res 52(15):4412–4423
8. Mabkhot M (2018) Requirements of the smart factory system: a survey and perspective
9. Cline G (2017) Industry 4.0 and industrial IoT in manufacturing: a sneak peek
10. Vila C (2015) An approach to sustainable product lifecycle management (Green PLM). Procedia
Eng 132:585–592
11. Xin Y (2019) Dealing with knowledge management practices in different product lifecycle
phases within product-service systems 83:111–117
12. Alihassansodhro. Convergence of IoT and product lifecycle management in medical healthcare
(2018)
13. Koshechkin K. Implementation of digital technologies in pharmaceutical products lifecycle
14. Leal AG, Santiago A (2014) Integrated environment for testing IoT and RFID technologies
applied on intelligent transportation system in Brazilian scenarios
ECC Image Encryption Using Matlab
Simulink Blockset
1 Introduction
Cryptography is a science that ensures the security of sharing and storage of data by
individuals in the presence of adversaries.
This is done through techniques that ensure the confidentiality, authenticity and
integrity of the data.
As a science of secrecy, cryptology is divided into two; cryptography which aims
to build and prove Cryptosystems and cryptanalysis whose goal is to "break" the
systems built by cryptographers.
For a long time and in all previous eras, secret codes were the basis of cryp-
tography, until 1970, the development of an encryption and signature system was
an essential objective of cryptography [3, 9, 10], therefore cryptanalysts wage
their wars in the shadows by controlling communication networks and information
transformation.
S. Chillali (B)
Sidi Mohamed Ben Abdellah University, FP, LSI, Taza, Morocco
e-mail: [email protected]
L. Oughdir
Sidi Mohamed Ben Abdellah University, ENSAF, Fez, Morocco
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 835
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_76
836 S. Chillali and L. Oughdir
The emergence of digital information sent through insecure channels like the
Internet has facilitated communication between different parties, in fact, there are
security holes in networks, which facilitates access to information.
Mathematics with difficult inverse problems and a support of computer science
and applied physics gave a power to cryptography which became a science attached
to mathematics and computer science; it is a science of encrypting messages, other
than cryptanalysis, which is used to decrypt them. [7, 8].
Our contribution in this work is the encrypted of different types of images of
different sizes, using a mathematical object called elliptic curve, or we can give
several definitions, studied in algebraic geometry, [4].
For more than forty years Diffie-Hellman, [1] gave the appearance of the key
exchange protocol which was developed at the end of the 1980s: ECC, Elliptic Curve
Cryptography. These cryptographic protocols use group structures whose security is
ensured by the difficulty of the discrete logarithm problem on these groups, (DLP),
which shows the success of these elliptic curves in public key cryptographic systems,
hence the interest of the arithmetic study of these geometric curves.
Image processing is certainly the most innovative that man has known, [5, 6] moti-
vated by his needs in various fields, including imaging and the following, generates
sensitive problems solved by the various image processing techniques.
The digital image is represented by a matrix of points called pixels (Picture
Element), each having as characteristic a gray level or a color coming from the
corresponding location in the real image or calculated from an internal description
of the scene to represent [2].
We can represent an image by a matrix M(mi,j ); mi,j ∈ {0, 1, 2, . . . , 255}.
ECC Image Encryption Using Matlab Simulink Blockset 837
In [11], the authors gave a new image encryption algorithm based on a secure
variant of Hill Cipher (HC) and three enhanced one dimensional chaotic maps (1D).
Our approach is to propose an encryption scheme based on a safer concept, namely
discrete logarithm problem on elliptic curves. Thus, the simulation results carried
out on a database of images in colors and in grayscale give very effective security
results, this is the subject of an article already submitted.
In the following, thanks to the discrete logarithm problem of this elliptic curve,
a new method of image encryption will be constructed, using the Diffie-Hellman
protocol on such a curve, the security of this type of encryption is proven because
the resolution of this problem is difficult on these curves (Fig. 1).
2 Elliptic Curve on F P
y 2 = x 3 + ax + b, (1)
kP = P
+P+
· · · + P (3)
k times
838 S. Chillali and L. Oughdir
Ea,b (p), + is an abelian group, [0:1:0] is the point at infinity it is the neutral
element for the + law.
Let Q also be an element of a cyclic subgroup G =< P >. An integer k that
solves the equation kP = Q is called a discrete logarithm of Q in the base P. One
writes k = logP Q.
Let P and Q, find x satisfying xP = Q, is called discret logaritm problem (DLP)
(Fig. 2).
Let p be a large prime number, Ea,b (p) is an elliptic curve on the field FP defined by
equation (1), whose discrete logarithm problem is difficult in this curve.
Aicha and Omar publish a point P which generates a subgroup of known order r of
an elliptic curve Ea,b (p).
• Aicha chooses a private number t ∈ {1, 2, . . . , r − 1} and calculates tP.
• Omar chooses a private number s ∈ {1, 2, . . . , r − 1} and calculates sP.
• Aicha sends tP to Omar.
ECC Image Encryption Using Matlab Simulink Blockset 839
Each for him, the two entities build the secret encryption matrix T by the following
two steps:
• Step1:
Build a matrix A = (ai,j )i=1,...,f;j=1,2,3 .
For 1 ≤ i ≤ f , if i K = (xi , yi ), then ai,1 = i, ai,2 = xi and ai,3 = yi , i.e.:
⎛ ⎞
1 x1 y1
⎜2 y2 ⎟
⎜ x2 ⎟
⎜ ⎟
⎜: : : ⎟
A=⎜ ⎟
⎜i xi yi ⎟
⎜ ⎟
⎝: : : ⎠
f xf yf
• Step2:
Build a Key matrix T that is the same size as the image.
One of the algorithms at this work is to find the key matrix T from matrix A.
Let [n, m, p] the size of the image to be encrypted, then T (Fig. 3).
• Flowchart Algorithm:
4 Encryption Algorithm
Aicha transforms R into a matrix Rt and then encrypt the image “encryptimage”
by the following encryption system:
5 Decryption Algorithm
Omar receive, "encryptimage" sent by Aicha, he has the private key which allows him
to calculate the decryption function and to find the matrix M, which can transform
it into the image, “image”, see Fig. 9.
We can summarize the encryption and decryption method by a flowchart see
Fig. 10.
842 S. Chillali and L. Oughdir
So;
E 1758,254 (2543) = (x, y) ∈ F22543 /y 2 = x 3 + 1758x + 254 ∪ {[0 : 1 : 0]}
ECC Image Encryption Using Matlab Simulink Blockset 843
• Image Entropy.
This study is a very important task.
Such an evaluation of these encryption algorithms proposed in this work was
carried in an article entitled: “Image encryption algorithm based on elliptic curves”
and submitted for publication in a journal Soft Computing.
7 Conclusion
Acknowledgments The authors thank the reviewers and everyone who contributed to this work.
References
1. Diffie W, Hellman M (1976) New directions in cryptography. IEEE Trans Inf Theory
2. Chillali S, Oughdir L (2019) A diagram of confidentiality of information during a traffic offence.
AIP Conf Proc 2074:020028
3. Zeriouh M, Chillali A, Boua A (2019) Cryptography based on the matrices. Bol Soc Paran Mat
37(3):75–83
846 S. Chillali and L. Oughdir
4. Boulbot A, Chillali A, Mouhib A (2020) Elliptic curves over the ring R. Bol Soc Paran Mat
38(3):193–201
5. Hua ZY, Zhou YC, Pun CM, Chen CLP (2015) 2D sine logistic modulation map for image
encryption. Inf Sci 297:80–94
6. Zhang Y (2018) The unified image encryption algorithm based on chaos and cubic. S-Box Inf
Sci 450:361–377
7. Zhang L, Zhang F (2009) A new certificate less aggregate signature scheme. Comput Commun
32(6):1079–1085
8. Xiong H, Guan Z, Chen Z, Li F (2014) An efficient certificate less aggregate signature with
constant pairing computation. Inform Sci 219:225–235
9. He D, Chen J, Zhang R (2012) An efficient and provably-secure certificate less signature scheme
without bilinear pairings. Int J Commun Syst 25:1432–1442
10. Karati A, Islam SKH, Biswas GP (2018) A pairing-free and provably secure certificate less
signature scheme. Inf Sci 450:378–391 (2018)
11. Essaid M, Akharraz I, Saaidi A, Mouhib A (2019) Image encryption scheme based on a new
secure variant of Hill cipher and 1D chaotic maps. J Inf Secur Appl 47:173–187
Development of Large Chaotic S-boxes
for Image Encryption
Abstract In this work, we suggest a novel algorithm for construction of large chaotic
confusion and diffusion S-boxes for gray scale and color image scrambling. The
modification of the value of a pixel by these S-boxes, depend not only on its gray
level intensity but also on its position in the original image. In order to raise the
impact of avalanche effect and to make our method robust against differential attack,
we have developed a strong link between the ciphered pixel and the next original.
The simulations carried out by our system on a large number of randomly selected
images show the effectiveness of this crypto system against any known attack.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 847
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_77
848 Y. Qobbi et al.
2 Proposed Method
Our technique is based on the use of three large chaotic S-boxes generated from
the chaotic maps to establish substitution and diffusion. The substitution of a pixel
depends not only on its value but also on its position. This technique is based on the
following steps:
• Preparation of Plain Image.
• Generation of Chaotic Sequences.
• Encryption Settings.
• Encryption Process.
• Decryption Process.
After loading of the plain image of size h × w: where h and w present the height and
width of plain image respectively. This image is subdivided into three vectors VR,
VG and VB of size 1 × t : where t = h × w. which are represents the red, green and
blue channels respectively. This transformation is illustrated in the Fig. 1.
For the generation of encryption keys, our method is based on the use of chaotic
dynamic systems. Indeed we use two chaotic maps.
Logistic Map. Is a chaotic dynamic system most used for the generation of pseudo-
random number sequences. It is given by the following expression:
Tent Map. It is the simple chaotic dynamic system used in image encryption. It is
described by the recursion relation:
µ2 yn i f yn < 0.5
yn+1 = : µ2 ∈ [0, 2] and y0 ∈ ]0, 1]. (2)
µ2 (1 − yn )Other wise
Our approach is based on the use of three S-boxes (SR, SG and SB) of size t × 256.
Where t = h × w. Which are used to encrypt the channels (Red, Green and Blue).
Generating technique of three s-boxes is similar. In the next section, we present the
method of generating the one of the three S-boxes dented SX. Where SX represents
SR, SG or SB.
Step 1: the initial conditions and the control parameters are set for the logistic
and tent maps.
Step2: the logistic and tent maps are iterated for 3 × 256 times to generate two
random sequences X and Y of size (3 × 256).
Step 3: The two sequences X and Y are combined to generate three chaotic vectors
P, Q and R of size (1, 256) by using the following algorithm:
P = (int)((max(X(i),Y(i)) × 1012 ) i = 0 to 255
Q = (int)((max(X(i),Y(i)) × 1012 ) i = 256 to 2 × 256–1
R = (int)((max(X(i),Y(i)) × 1012 ) i = 512 to 3 × 256–1
Step 4: By sorting in ascending order the elements of the vectors P, Q and R, three
permutation vectors U, V and W of size (1, 256) are obtained.
Step 5: the elements of the vectors U, V and W, which contain the values
in [[0, 255]] are used to initialize the three first rows of SX. The other rows are
generated by the following algorithm:
SX(i,j) = SX(i-2 , SX(i-1,j)) i = 3 to t-1 and j = 0
to 255
Steps 1 to 5 are repeated three times to generate SR, SG and SB.
Development of Large Chaotic S-boxes for Image Encryption 851
After decomposing the Plain image into three vectors VR, VG and VB and generation
the encryption parameters, the proposed encryption process is applied according to
the following steps:
Step 1: Modification of the first pixel of each vector VR, VG, and VB by using
the initialization vector (Iv) to start the diffusion process.
Step 2: The three S-boxes (SR), (SG) and (SB) are used to encrypt the pixels of
three color vectors VR, VG and VB. This encryption of a pixel depends not only on
its value but also on its position in the original image, using the following algorithm:
VR(i) = SR (i,VR(i)) where i = 0 to t-1
VG(i) = SG (i,VG(i)) where i = 0 to t-1
VB(i) = SB (i,VB(i)) where i = 0 to t-1
Step 3: In order to establish a diffusion process, the encrypted pixel is linked to
the next original pixel, by using the strong expression as follows:
VR(i + 1) = SB(i, VR(i)) ⊕ VR(i + 1) where i = 0 to t-2
VG(i + 1) = SR(i, VG(i)) ⊕ VG(i + 1) where i = 0 to t-2
VB(i + 1) = SG(i, VB(i)) ⊕ VB(i +1) where i =0 to t-2
Where ⊕ is the Xor operation.
The diagram below (Fig. 2) represents the encryption mechanism of the suggested
scheme.
Our approach is based on diffusion process, which requires starting the decryption
process from the last pixel to the first. This process goes through the following steps:
Step 1: split the cipher image of size (h, w) into three vectors VR, VG and VB
of size (1, t) where t = h × w.
852 Y. Qobbi et al.
Step 3: The three large invers S-boxes (RS, G S and B S) are used to find the
original pixel by using the following algorithm:
Development of Large Chaotic S-boxes for Image Encryption 853
3 Experimental Results
To prove the effectiveness of our scheme against any known attack. We present
several simulation results obtained by application of our crypto system on Lena (256
× 256), Peppers (512 × 512), and Baboon (512 × 512), The results are given in the
following diagram (Fig. 3).
4 Security Analysis
The main objective of any encryption system is to secure the information from any
known attack. Some security analysis has been applied on the proposed algorithm
which includes correlation, histogram, key space, sensitivity and entropy analysis.
The result analysis of our system is described as follows.
The adjacent pixels in data image are generally highly correlated. So an efficient
scheme is the one that allows removing this correlation to resist against statistical
attacks. The correlation between adjacent pixels is given by the following equation:
1 N 1 N
E(x) = xi and E(y) = yi (4)
N 1 N 1
1 N 1 N
D(x) = (xi − E(x))2 and D(y) = (yi − E(y))2 (5)
N 1 N 1
Where Pr(mi ) is the probability of random variable m of ith index.The ideal entropy
for 256 symbols representing the different gray levels in an 8-bit image is H(m) = 8.
856 Y. Qobbi et al.
The entropy of the different ciphered images obtained by our scheme is near to 8.
That prove our scheme can resist against the attacks by entropy. The information
entropy of our algorithm and some references are listed in the Table 2 below.
Key Space
A good encryption system should have a key space large enough to remain strong
against brute force attacks. Indeed the key space must be greater than 2100 [13].The
key space of our method is composed of two initial conditions x0 and y0 and two
control parameters µ1 and µ2 . These parameters are the float (32 bit). Then the key
space of our system is 2128 . This makes our system safe to any brute force attack.
A strong encryption system is one that is sensitive to a small change in the original
image such as a single pixel change in the original image produces entirely different
ciphered image. This sensitivity is evaluated through a Number of Pixels Change
Rate (NPCR) and Unified Average Changing Intensity (UACI). Their definitions are
presented as follows.
w h
1 Di j
N PC R = 1
× 100% (7)
w∗h
1 i j I C1i j − I C2i j
U AC I = × 100% (8)
w×h 255
IC1 = [IC1i j ] and IC2 = [IC2i j ] express two ciphered images corresponding to
two plain images with a tiny difference, and Di j is calculated by
1 i f I C1i j = I C2i j
Di j = (9)
0 else
Development of Large Chaotic S-boxes for Image Encryption 857
The following Table 3 below shows the measurement of NPCR and UACI between
two cipher images of the Babon (512 × 512) Lena (256 × 256) and Peppers (512 ×
512). When making a slight modification in the original image.
5 Conclusion
In this article, a novel technique is developed to create three chaotic S-boxes. Which
are used to perform two main processes of any image encryption system: Substitution
and diffusion. The main contributions of our approach are: First, the modification of a
value of a pixel depends on two parameters which are: its value and its position in the
plain image. Second, our algorithm uses three large chaotic S-boxes of dynamic size.
The simulation results and security analysis presented by a correlation coefficient
close to zero, a flat histogram of the ciphered image and an entropy value near to 8,
prove that our algorithm is strong against any known attack.
References
1. Zhang Y (2018) The unified image encryption algorithm based on chaos and cubic S-Box. Inf
Sci 450:361–377
2. Fridrich J (1998) Symmetric ciphers based on two-dimensional chaotic maps. Int J Bifurcat
Chaos 8(06):1259–1284
3. Cao C, Sun K, Liu W (2018) A novel bit-level image encryption algorithm based on 2D-LICM
hyperchaotic map. Signal Process 143:122–133
4. Essaid M, Akharraz I, Saaidi A (2019) Image encryption scheme based on a new secure variant
of Hill cipher and 1D chaotic maps. J Inf Secur Appl 47:173–187
5. Hraoui S, Gmira F, Abbou MF, Oulidi AJ, Jarjar A (2019) A new cryptosystem of color image
using a dynamic-chaos hill cipher algorithm. Procedia Comput Sci 148:399–408
6. Belazi A, Abd El-Latif AA (2017) A simple yet efficient S-box method based on chaotic sine
map. Optik 130:1438–1444
7. Wang X, Wang Q (2014) A novel image encryption algorithm based on dynamic S-boxes
constructed by chaos. Nonlinear Dyn 75(3):567–576
8. Zhu S, Wang G, Zhu C (2019) A secure and fast image encryption scheme based on double
chaotic s-boxes. Entropy 21(8):790
9. Khan M (2015) A novel image encryption scheme based on multiple chaotic S-boxes. Nonlinear
Dyn 82(1–2):527–533
858 Y. Qobbi et al.
10. Lu Q, Zhu C, Deng X (2020) An efficient image encryption scheme based on the LSS chaotic
map and single S-box. IEEE Access 8:25664–25678
11. Zahid AH, Arshad MJ (2019) An innovative design of substitution-boxes using cubic
polynomial mapping. Symmetry 11(3):437
12. Rehman AU, Khan JS, Ahmad J, Hwang SO (2016) A new image encryption scheme based on
dynamic s-boxes and chaotic maps. 3D Res 7(1), 7.
13. Wu X, Zhu B, Hu Y, Ran Y (2017) A novel color image encryption scheme using rectangular
transform-enhanced chaotic tent maps. IEEE Access 5:6429–6436
Evaluation of Feature Extraction
Methods Combined with Support Vector
Machines for Powerline Component
Recognition in Aerial Images
Abstract As part of its efforts to maintain the stability and sustainability of power
generation for its consumers, the Moroccan National Office of Electricity and
Drinking Water (ONEE) conducts regular inspections of its transmission and distri-
bution networks. Compared to other inspection methods, UAV inspection enables
regular inspection with minimized effort and lower cost and provides more detailed
images of network components safely and securely. However, human analysis of the
amount of data produced by the UAV may not be efficient. Therefore, the image
analysis process’s automation should be implemented by developing advanced algo-
rithms to detect components and their defects. In this work, we focused on object
recognition, which we consider a key step in the development of more complex tasks.
We evaluated HOG, LBP, and SIFT extractors’ recognition performance in the iden-
tification of power line components, by combining each of these extractors with the
SVM classifier. The results showed that combining the HOG and SVM can give the
best results compared to the LBP-SVM and SIFT-SVM methods with an accuracy
rate of 97.98% and a shorter acquisition time for image processing. These results
indicate that this combination can meet electric companies’ technical requirements
related to accuracy and speed.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 859
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_78
860 J. Garfaf et al.
the pillar of the energy strategy and the armed arm of the state in the water and sanita-
tion sector in Morocco [1], conducts regular diagnoses and visual inspection of their
transmission and distribution networks. These inspection missions within the ONEE
were until recently carried out by patrols on the ground and by helicopters. However,
these traditional methods have several drawbacks. In addition to the fact that these
methods are expensive and too slow, they jeopardize the safety of individuals given
that most components are located in hard-to-reach places and that these inspection
operations are often carried out under more difficult meteorological conditions [2, 3].
Accordingly, the ONEE devotes significant investment and effort to carry out these
types of operations.
To ensure inspection and maintenance operations with the lowest possible cost
and effort as well as ensure the safety of agents and operators, the ONEE has resorted
to the implementation of UAV technology at its transmission and production units.
Compared to other inspection methods based on foot patrols and helicopter-assisted
inspections, operating costs are lower, and the UAVs can fly safely as close as possible
to the power line element, allowing more detailed photos of the components to be
taken [3–6]. However, given the enormous amount of data produced by the UAV,
manual human analysis can be ineffective. This explains the use of artificial intelli-
gence as an engineering solution to replace humans and optimize results, as is the
case for many other engineering problems [7]. Indeed, it is necessary to develop effi-
cient and advanced algorithms to automate the process of identifying the presence
of anomalies or damage on inspected power lines by exploiting the aerial vector of
photos and videos taken and produced remotely by the UAV. Numerous studies are
being conducted on the monitoring and inspection of power transmission systems.
In this work, we are focusing on the detection and recognition of power network
components as an essential part of the development of complex algorithms.
In [8], a supervised learning approach has been proposed to detect and classify
pylons. This approach uses Histogram of Oriented Gradients (HOGs) to train two
MLP (Multilayer Perceptron) neural networks. The first classification separates the
foreground from the background, and the second multilayer MLP distinguishes the
four different types of electrical pylons described in the paper. The same feature
extractor has been used in [9] for the detection of molten insulators by the local
binary model (LBP). The extracted features were transferred to the SVM (Support
Vector Machine) classifier. An evaluation of the proposed method was performed on
500 images, which gave a detection rate of 89.1%.
Using the infrared images introduced by the UAV, ZHOU et al. [10] designed a
system for fault diagnosis and remote isolator detection. Each isolator is located using
the Scale-invariant Feature Transform (SIFT) method. The obtained results showed
that the choice of SIFT allows the identification and localization of isolators with a
high accuracy. This boost is thanks to the invariance of scale, and to the invariance
to variations in rotation angle and image brightness. Because of its performance, the
features of SIFT are also used by [11]. They are mainly used to detect the tilt angle
of the electric pylon during video surveillance in the substation.
Among the most prominent traditional object extractors, are: Scale-Invariant
Features Transform (SIFT) [12], Histogram of Oriented Gradients (HOG) [13], Local
Evaluation of Feature Extraction Methods … 861
Binary Patterns (LBP) [14] which have been used in many works, we can cite among
others [15–17]. In order to evaluate their recognition performance in the identifica-
tion of power line components, each of these extractors will be combined with the
famous SVM classifier which has proven in many works its performance in computer
vision applications. The experiments were conducted on a set of images of power
line components captured from videos taken by the UAV.
We organized this paper as follows: the second section describes in detail the
methodology adopted and the algorithms studied and used. The results are presented
and discussed in the third section. The last section concludes the paper while
highlighting future research directions.
2 Methodology
In the computer vision domain, object recognition is a crucial and challenging task.
It consists of determining the predefined category to which an object in a given
image belongs [18]. The recognition process consists of two main steps: the feature
extraction step, followed by the classification step [19]. Indeed, before the received
image is fed into the recognition process, a pre-processing phase is necessary to
improve the quality of the input images so that the relevant features can be better
extracted from the image. This preprocessing phase consists firstly of converting the
images produced by the UAV, which are generally in RGB colors, into grayscale
[20, 21] and secondly of resizing them to a smaller fixed size of 256 * 256 so that
no important information is missing. This allows to prevent the overloading and
saturation of the network and to reduce the computing time. Then, a feature extractor
is placed whose objective is to represent the recovered image as a descriptive vector
based on the extracted features. This vector is afterwards passed to the classifier to
recognize the class to which the image belongs (see Fig 1).
For an image recognition system to be effective, it must have a well-defined
and highly targeted feature extraction step. To achieve this, the features must be as
distinctive and informative as possible in order to obtain reliable identification. In fact,
if the features are not well selected, this could significantly impair the performance
and efficiency of the recognition system.
Among the most popular traditional object extractors are SIFT, HOG, LBP, which
have been used in many works. These extractors will be the main subject of this paper.
An evaluation of their recognition performance will be performed in our paper by
combining them with the famous SVM classifier, which is becoming more and more
popular [22, 23] due to its performance in the classification of noisy and high-
dimensional data [24]. The experiments are conducted on a data set constructed
from aerial images of power line components.
A detailed description of the HOG, LBP and SIFT extractors and the SVM
classifier is given in the following sections.
The gradient histogram was proposed initially by Dalal and Triggs [13] for human
detection. The gradient-oriented histogram is used to extract features using informa-
tion about the orientation of the gradient from the appearance and shape of the local
object.
The calculation of HOG descriptors is based on four key operations: calculation
of the amplitude and orientation of the image gradient, generation of cell histograms,
histogram normalization and descriptor generation. First, a detection window of 64
* 128 pixels is divided by cells of 8 * 8 pixels, forming 8 * 16 = 128 cells. For
each pixel (x, y), the gradient magnitude m(x, y) and the orientation θ(x, y) can be
calculated using Eqs. (1) and (2), respectively:
m(x, y) = G x (x, y)2 + G y (x, y)2 (1)
G y (x, y)
(x, y) = arctan (2)
G x (x, y)
where G x (x, y)et G y (x, y) represent the gradient of the image along the two
directions x and y successively as given below:
G x (x, y) = I (x + 1, y) − I (x − 1, y) (3)
With I (x, y) represent the intensity of the image in the pixel (x,y). Then, a
histogram is generated for each cell containing nine boxes. The amplitude of each
cell in the histogram is calculated from the accumulation of the amplitudes of the
pixels with directions belonging to the interval of the cell. A 9 × 1 vector is then
obtained for each cell and presented as a histogram. The blocks are normalized to
give a 36 * 1 vector in order to have a better invariance to changes in illumination and
contrast. The process is repeated by passing through all the pixels of the image. In
the end, all 36 × 1 normalized vector is grouped into a single vector corresponding
to the HOG descriptor.
Evaluation of Feature Extraction Methods … 863
Local Binary Pattern (LBP), originally proposed by Ojala in 1996 [14], allows char-
acterizing the spatial structure of the image texture using a 3 * 3 neighborhood around
a central pixel. Based on a 3 * 3 neighborhood, the LBP compares the gray values
of the 8 neighboring pixels to the central pixel, which is taken into account as a
threshold.
Let gc , g0 ,…, g7 denote respectively the grey levels of the central pixel and its
eight neighboring pixels. The LBP code of the central pixel characterized by its
coordinates (xc , yc ) is calculated as follows:
7
L B P(xc , yc ) = S(gc − gi ) × 2 p (5)
p=0
The Fig 2 shows an example for the calculation of the LBP code
After generating the LBP codes of all pixels of the input image, these codes are
collected in a histogram in the form of a texture descriptor.
Scale-Space Extrema Detection: :This step aims to scroll through the image at
different scales and in different octaves in order to extract the points of interest that
864 J. Garfaf et al.
are invariant to the scale. For this, the input image is convoluted with a centered
Gaussian with a progressive variation of σ. A succession of images with different
levels of blur is obtained, thus constructing a Gaussian pyramid. The points of interest
are considered as the local maxima and minima in DoG images that appear at different
scales through the application of the Gaussian difference between two successive
images of the same scale.
Keypoints Localization: The first step of the SIFT algorithm produces a vast
number of points of interest, some of which are unstable and cannot be retained as
characteristics. An adjustment is, therefore, necessary to obtain more useful points.
Low contrast points and points located on low curvature contour edges are therefore
eliminated in this step.
Orientation Assignment: In this step, each selected point is assigned one or more
orientations based on the local features of the image. Indeed, for each point of interest,
the algorithm takes a neighborhood, then calculates the amplitude and direction of
the gradient at each pixel to finally build an orientation histogram with 36 graduations
covering 360°. The peaks of the histogram characterize the dominant orientation in
the neighborhood of a point of interest. In order to have more information available,
Lowe suggests using the peaks of the histogram that represent at least 80% of the
dominant peak to create new points of interest that differ in their main orientation.
A set of key points that are invariant in scale and rotation is then created.
Keypoints Descriptor: the descriptors are calculated according to the orientations
of the key points. Indeed, the algorithm defines a neighborhood window of 16 * 16
pixels around the point of interest. This window is divided into 16 blocks of 4 *
4 pixels where the orientation and norm of the gradients are calculated for each of
these blocks. These orientations are then placed in a histogram with 8 graduations
and are weighted by the norm of the gradient at the point of interest to reduce the
illumination change effects. The result is a SIFT feature vector of 16 ∗ 8 = 128
values for each point of interest.
SVM was originally developed by Vapnik [26] for two-class classification problems
but was later extended to include multi-class classification problems. SVM aims to
define decision boundaries by providing hyperplanes to classify data samples in data
distinctly. The data of the input space is projected into a larger space using the kernel
function [24].
For a training set labeled (xi , yi ), i = 1, 2… where xi Rn and yi {1, −1}, the
SVM proceeds to solve the optimization problem that may be presented as follows:
WTW
Min w,b,ξ +C ξi (7)
2 i
Evaluation of Feature Extraction Methods … 865
Subject to
yi (w T φ(xi ) + b) ≥ 1 − ξi , ξi ≥ 0 (8)
where w is the weight vector, and b is the bias. The training vector xi is mapped to a
higher dimensional space by the function φ, and the coefficient C is set to adjust the
error tolerance.
Several kernel functions are applied according to the SVM method. The RBF
kernel is used in our paper, whose function is as follows:
−γ xi −x 2j
K xi , x j = e ,γ > 0 (9)
0
HOG-SVM SIFT-SVM LBP-SVM
precision, recall, F1-score, and accuracy which are defined as follows [29]:
TP
Pr ecision = (10)
T P + FP
TP
Recall = (11)
T P + FN
2T P
F1 − scor e = (12)
2T P + T N + F P + F N
TP +TN
Accuracy = (13)
T P + T N + FP + FN
Where TP, FP, TN and FN are the numbers of true positives, false positives, true
negatives, and false negatives in the detection results, respectively.
As shown in Fig 4, it can be seen that the HOG-SVM model achieves the best
performance in power line recognition with 97.98% accuracy, compared to SIFT-
SVM and LBP-SVM, which achieved an accuracy of 45.78% and 32.37%, respec-
tively. In addition, Table 3 shows the other evaluation parameters related to the
Evaluation of Feature Extraction Methods … 867
accuracy, recall and f1 scores. It shows that the combination of HOG and SVM
presents a good compromise between accuracy and recall. Recall reaches 97.96%,
indicating that most components are correctly detected.
Besides, given that the solution will be implemented in real-time applications [30],
the image processing time for each model must also be calculated and compared. As
shown in Table 4, the image processing time for the HOG-SVM model is 0.349 s,
which is less than the other two models. Both the SIFT-SVM model and the LBP-
SVM are very time-consuming, resulting from the large amount of data and the high
resolution of the collected images.
Experimental results, based on the comparative study and evaluation of models
designed using different evaluation measures, show that the method in which HOG
is the extractor combined with SVM gives the best results in terms of accuracy
and computing time. These results indicate that this combination allows utilities to
meet their technical requirements for accuracy and speed when performing off-line
analysis for transmission line inspection.
4 Conclusion
Acknowledgement This work was supported by the National Center for Scientific and Technical
Research (CNRST, Morocco) and the National Office for Electricity and Drinking Water (ONEE,
868 J. Garfaf et al.
Morocco). We want to express our gratitude to the ONEE teams for their collaboration and support
for successfully conducting our research project.
References
1. The official Website of the ONEE - Electricity sector (2020). http://www.one.org.ma/. Accessed
26 Oct 2020
2. Nguyen VN, Jenssen R, Roverso D (2018) Automatic autonomous vision-based power line
inspection: a review of current status and the potential role of deep learning. Int J Electric
Power Energy Syst 99: 107–120
3. Fahmani L, Garfaf J, Boukhdir K, Benhadou S, Medromi H (2020) Modelling of very high
voltage transmission lines inspection’s quadrotor. SN Appl Sci 2: 1425
4. Liu X, Miao X, Jiang H, Chen J (2020) Review of data analysis in vision inspection of power
lines with an in-depth discussion of deep learning technology. arXiv:2003.09802[cs,eess]
5. Jalil B, Leone G, Martinelli M, Moroni D, Pascali M, Berton A (2019) Fault detection in power
equipment via an unmanned aerial system using multi modal data. Sensors 19: 3014
6. Fahmani L, Garfaf J, Boukhdir K, Benhadou S, Medromi H (2020) Unmanned aerial vehicles
inspection for overhead high voltage transmission lines. In: 2020 1st International Conference
on Innovative Research in Applied Science, Engineering and Technology (IRASET). pp 1–7.
IEEE, Meknes, Morocco
7. Chergui M, Chakir A, Medromi H (2019) Smart IT governance, risk and compliance semantic
model: business driven architecture. In: 2019 Third World Conference on Smart Trends in
Systems Security and Sustainablity (WorldS4), pp 297–301. IEEE, London, United Kingdom
8. Sampedro C, Martinez C, Chauhan A, Campoy P (2014) A supervised approach to electric tower
detection and classification for power line inspection. In: 2014 International Joint Conference
on Neural Networks (IJCNN), pp 1970–1977. IEEE, Beijing, China
9. Tiantian Y, Guodong Y, Junzhi Y (2017) Feature fusion based insulator detection for aerial
inspection. In: 2017 36th Chinese Control Conference (CCC), pp 10972–10977. IEEE, Dalian,
China
10. Shen-pei Z, Xi L, Bing-chen Q, Hui H (2017) Research on insulator fault diagnosis and remote
monitoring system based on infrared images. Procedia Comput Sci 109: 1194–1199
11. Yu P, Dong BG, Xue YJ (2012) Electric power tower inclination angle detection method based
on SIFT feature matching. AMM. 236–237: 759–764
12. Lowe DG (1999) Object recognition from local scale-invariant features. In: Proceedings of
the Seventh IEEE International Conference on Computer Vision, pp 1150–1157, vol 2. IEEE,
Kerkyra, Greece. https://doi.org/10.1109/ICCV.1999.790410
13. Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: 2005 IEEE
Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2005), pp
886–893. IEEE, San Diego, CA, USA
14. Ojala T, Pietikäinen M, Harwood D (1996) A comparative study of texture measures with
classification based on featured distributions. Pattern Recogn 29: 51–59
15. Jones MJ, Viola P (2010) Robust Real-time Object Detection, 30
16. Taşkiran M, Çam ZG (2017) Offline signature identification via HOG features and artifi-
cial neural networks. In: 2017 IEEE 15th International Symposium on Applied Machine
Intelligence and Informatics (SAMI), pp 000083–000086
17. Zhang H, Qu Z, Yuan L, Li G (2017) A face recognition method based on LBP feature for CNN.
In: 2017 IEEE 2nd Advanced Information Technology, Electronic and Automation Control
Conference (IAEAC), pp 544–547
18. Liu L, Ouyang W, Wang X, Fieguth P, Chen J, Liu X, Pietikäinen M (2019) Deep Learning for
Generic Object Detection: A Survey. arXiv:1809.02165[cs]
Evaluation of Feature Extraction Methods … 869
Abstract Segmentation has a very essential role in medical imaging. Image segmen-
tation can be defined as the process that gives the ability to separate a random image
into parts or objects that make up the image, which means separating objects from
the background, so we can later analyze each of these components separately. In
this work we’ll propose a method based on the combination between fuzzy c-means
clustering and marker watershed algorithm. The first thing to do is to apply the fuzzy
c-means clustering which is an unsupervised clustering method that is usually used
in medical imaging. Then the first output image will be going through the second
algorithm which is the marker watershed that uses concepts from edge detection
and mathematical morphology. The proposed method was applied and tested on
several medical images to detect anomalies such as tumors, and according to the
results, we can say that it gives a satisfactory and good results of detection. To eval-
uate the performance of our proposed method, we used a set of metrics such as the
“Dice coefficient”, “Sensitivity” and “Specificity”; which all have shown satisfactory
results.
1 Introduction
When we talk about Medical imaging we’ll be definitely looking for a technique or
a process of making the interior of the body visually represented. Medical imaging
is used for clinical analysis and medical intervention, such as revealing any structure
hidden by bones or skin. It’s also very helpful when doctors try to diagnose and treat
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 871
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_79
872 H. Moussaoui et al.
2 Related Works
In the past 10 years many brain tumor detection methods have been proposed, we
can start by fuzzy c-means algorithm which plays a very important role in medical
image segmentation, nevertheless it has many drawbacks. By using fuzzy c-means
algorithm we need first to set the number of clusters which is considered as a disad-
vantage, and also it’s known with it sensitiveness about noise and other artifacts. M.
Sreerangappa, M. Suresh and D. Jayadevappa [17] have proposed a modified fuzzy
c-means algorithm known as a spatial FCM clustering to avoid all the drawbacks
listed above. For this methodology they started by pre-processing the MRI image by
applying the Median Filter followed by Wavelets Transform and Spatial FCM. Nilesh
Bhaskarrao Bahadure, Arun Kumar Ray and Har Pal Thethi [18] have proposed a
combination between Berkeley Wavelet Transformation (BWT) and SVM that helps
to classify MRI images to healthy or infected tumor. P. Kumar and B. Vijayakumar
[19] have proposed a brain tumor segmentation and classification methodology, based
on Principal Component Analysis (PCA) and Radial Basis Function (RBF).
c
uj (xi ) = 1 i = 1, 2, . . . , k (1)
j=1
A Novel Brain Tumor Detection Approach Based on Fuzzy C-means … 873
Step (3): Calculate dissimilarity between the data points and centroid using the
Euclidean distance:
Di = (x2 − x1 )2 (y2 − y1 )2 (3)
Step (4): Upgrade the new membership matrix. For this we use the equation
bellow:
1/
1 m−1
d ji
u j (xi ) = (4)
c 1/
1 m−1
k=1 dki
Calculate”Euclidean distance”
No
Output
874 H. Moussaoui et al.
5 Thresholding
T
ω= P(i) when P(i) = n i /N (5)
i=0
T
μ= i P(i)/ω (6)
i =0
A Novel Brain Tumor Detection Approach Based on Fuzzy C-means … 875
Threshold segmentation
Local Global
”Bernsen”
thresholding thresholding
”Chow and Kanelo”
”Eikvil”
”Mardia and
Hainsworth”
Traditional Multistage Iterative
”Niblack”
(Otsu) (Quadratic ratio (iterative triclass
”Yanowitz and
technique) thresholding)
Bruckstein”
”Tr Singh”
”Sauvola”
T
σt2 = (i − μ)2 P(i) (7)
i=0
In this section we’ll present our proposed method which is made of two stages.
First of all we’ll start by applying fuzzy c-means clustering on our input image
which is supposed to be a MRI image, and then the first output image will be going
through the second algorithm which is the marker controlled watershed. The reason
behind choosing fuzzy c-means clustering wasn’t arbitrary; but, because of it robust
characteristics for ambiguity. This segmentation method is highly recommended
when we’re dealing with medical imaging, also, it’s a soft clustering method that
assigns a membership of every pixel to each cluster. The second stage is to apply
the watershed algorithm to the segmented image. The proposed method gives good
detection results. As we can see that the final result shows clearly the tumor area well
segmented and surrounded by a marker which defines the borders. The flow chart of
the proposed method is as follow:
876 H. Moussaoui et al.
Input image
First Stage
Applying fuzzy c-means segmentation
Watershed algorithm
Generate Markers
Final segmentation
In this experiment part we’ll be using brain web images that have a tumor. As we
can see in the flow chart in the top (Fig. 3), the proposed method is divided into two
stages. In the first stage we’ll apply fuzzy c-means algorithm to the input image,
here, we tried many values of k and the conclusion we can draw is that by choosing
a small value of k (k = 3) we will have good results at the end. After applying
fuzzy c-means we de-noise the first output image using Median filter. In the second
stage of the proposed algorithm, we’ll apply both thresholding and marker watershed
algorithm to identify and segment the tumor using markers in the brain image. As
we can see bellow in (Fig. 4), the proposed algorithm gives a good segmentation and
identification of tumors in different locations.
To evaluate the performance of the proposed method, we used a set of metrics
such as “Dice coefficient”, “Sensitivity” and “Specificity”.
Dice Similarity:
2 ∗ GT ∗ CS /|GT| + |CS| (8)
Fig. 4 A table showing the obtained results using the proposed method
Fig. 5 A table showing the obtained results for the first image
Specificity: an evaluating metric that indicates the true negativity, which means that
sensitivity gives us the probability that a segmented pixel doesn’t belong to the brain
tumor area, but belongs to the background.
TP: indicates the true positivity, which refers to the number of pixels that belong
to the tumor area.
TN: indicates the true negativity, which refers to the number of pixels that don’t
belong to the tumor area.
FP: The false positive, which indicates the number of pixels that were wrongly
detected as brain tumor pixels (Figs. 5 and 6).
878 H. Moussaoui et al.
8 Conclusion
In this paper we’ve been talking about our proposed method which is based on fuzzy
c-means clustering and marker controlled watershed algorithm. The results obtained
show satisfactory abnormalities detection. We’ve used several metrics to evaluate
the performance for the proposed method like “Dice coefficient”, “Sensitivity” and
“Specificity”.
References
1. Khrissi L, Elakkad N, Satori H, Satori K (2019) Image segmentation based on kmeans and
genetic algorithms. In: Proceedings of ESAI 2019, embedded systems and artificial intelligence,
pp 489–497 (2019)
2. Zheng X, Lei Q, Yao R, Gong Y, Yin Q (2018) Image segmentation based on adaptive K means
algorithm. EURASIP J Image Video Process 2018:1–10. https://doi.org/10.1186/s13640-018-
0309-3 Article number: 68
3. Zaitoun NM, Aqel MJ (2015) Survey on image segmentation techniques. In: International
conference on communication, management and information technology (ICCMIT 2015).
ScienceDirect
4. Ghosh S, Das N, Das I, Maulik U (2019) Understanding deep learning techniques for image
segmentation. ACM Comput Surv 52(4):1–35. https://doi.org/10.1145/3329784
5. El Akkad N, El Hazzat S, Saaidi A, Satori K (2016) Reconstruction of 3D scenes by camera
self-calibration and using genetic algorithms. 3D Res 6(7):1–17
6. Merras M, Saaidi A, El Akkad N, Satori K (2016) Multi-view 3D reconstruction and modeling
of the unknown 3D scenes using genetic algorithms. Soft Comput 22(19):6271–6289
7. El Akkad N, Merras M, Baataoui A, Saaidi A, Satori K (2017) Camera Self-calibration having
the varying parameters and based on homography of the plane at infinity. Multimed Tools Appl
77(11):14055–14075
8. El Akkad N, Merras M, Saaidi A, Satori K (2014) Camera self-calibration with varying intrinsic
parameters by an unknown three-dimensional scene. Vis Comput 30(5):519–530
9. El Akkad N, Merras M, Baataoui A, Saaidi A, Satori K (2018) Camera self-calibration having
the varying parameters and based on homography of the plane at infinity. Multimed Tools Appl
77(11):14055–14075
10. Akkad NE, Merras M, Saaidi A, Satori K (2013) Robust method for self-calibration of cameras
having the varying intrinsic parameters. J Theor Appl Inf Technol 50(1):57–67
11. Akkad NE, Merras M, Saaidi A, Satori K (2013) Camera self-calibration with varying
parameters from two views. WSEAS Trans Inf Sci Appl 10(11):356–367
A Novel Brain Tumor Detection Approach Based on Fuzzy C-means … 879
12. El Akkad N, Saaidi A, Satori K (2012) Self-calibration based on a circle of the cameras
having the varying intrinsic parameters. In: Proceedings of 2012 international conference on
multimedia computing and systems, ICMCS. pp 161–166
13. Merras M, El Akkad N, Saaidi A, Nazih AG, Satori K (2014) Camera calibration with varying
parameters based on improved genetic algorithm. WSEAS Trans Comput 13:129–137
14. Es-sabry M, El Akkad N, Merras M, Saaidi A, Satori K (2019) A novel color image encryption
approach based on random numbers generation of two matrices and bit-shift operators. Soft
Comput 24(5):3829–3848
15. Es-Sabry M, El Akkad N, Merras M, Saaidi A, Satori K (2018) A novel text encryption
algorithm based on the two-square cipher and caesar cipher. Commun Comput Inf Sci 872:78–
88
16. Es-Sabry M, El Akkad N, Merras M, Saaidi A (2018) Grayscale image encryption using shift
bits operations. In: 2018 international conference on intelligent systems and computer vision,
ISCV, pp 1–7
17. Sreerangappa M, Suresh M, Jayadevappa D (2019) Segmentation of brain tumor and perfor-
mance evaluation using spatial FCM and level set evolution. Open Biomed Eng J 13(1):M6.
https://doi.org/10.2174/1874120701913010134
18. Bahadure NB, Ray AK, Thethi HP (2017) Image analysis for MRI based brain tumor detection
and feature extraction using biologically inspired BWT and SVM. Int J Biomed Imaging
2017:12. Article ID 9749108. https://doi.org/10.1155/2017/9749108
19. Kumar P, Vijayakumar B (2015) Brain tumour Mr image segmentation and classification using
by PCA and RBF kernel based support vector machine. Middle-East J Sci Res 23(9):2106–2116
20. Maolood I, Al-Salhi Y, Lu S (2018) Thresholding for medical image segmentation for cancer
using fuzzy entropy with level set algorithm. Open Med J 13(1):374–383. https://doi.org/10.
1515/med-2018-0056
21. Jobin Christ MC, Parvathi RMS (2011) Fuzzy c-means algorithm for medical image
segmentation. https://doi.org/10.1109/ICECTECH.2011.5941851
22. Zhou H, Schaefer G, Shi C (2009) Fuzzy c-means techniques for medical image segmentation.
In: Jin Y, Wang L (eds) Fuzzy systems in bioinformatics and computational biology. Studies
in fuzziness and soft computing, vol 242. Springer, Heidelberg. https://doi.org/10.1007/978-3-
540-89968-6_13
23. Grau V, Mewes AUJ, Alcaniz M, Kikinis R, Warfield SK (2004) Improved watershed transform
for medical image segmentation using prior information. IEEE Trans Med Imaging 23(4):447–
458. https://doi.org/10.1109/TMI.2004.824224
24. Zhanpeng H, Qi Z, Shizhong J, Guohua C (2016) Medical image segmentation based on the
watersheds and regions merging. In: 2016 3rd international conference on information science
and control engineering (ICISCE), Beijing, pp 1011–1014. https://doi.org/10.1109/ICISCE.
2016.218
25. Lu Y, Jiang Z, Zhou T, Fu S (2019) An improved watershed segmentation algorithm of medical
tumor image. In: IOP conference series: materials science and engineering, vol 677, no 4
26. Ravi S, Khan AM (2014) Bio-medical image segmentation using marker controlled watershed
algorithm: a case study. Int J Res Eng Technol 3:26–30. eISSN: 2319-1163, pISSN: 2321-7308
27. Sandhya G, Kande GB, Savithri S (2017) A novel approach for the detection of tumor in
MR images of the brain and its classification via independent component analysis and kernel
support vector machine. Imaging Med 9(3):33–44 ISSN: 1755-5191
28. Senthilkumaran N, Vaithegi S (2016) Image segmentation by using Thresholding techniques
for medical images. Comput Sci Eng: Int J (CSEIJ) 6(1):1–13
Image Retrieval Based on MPEG-7
Feature Selection Using Meta-heuristic
Algorithms
Abstract The continuous growth of the internet and its diversified domains and
tools; and the rapid development of new technologies have enormously increased
the amount of digital images and consequently the dimension of the image data
sets. Therefore, finding the relevant images to a query image in such database is a
challenging task. Content-based image retrieval is the powerful method so far which
use the visual characteristics of the image. The choice of these visual descriptors is
a decisive phase in this area. This work presents a method which extracts the three
image features which are: color, texture and shape based on the MPEG-7 standards.
Then, the genetic algorithm is implemented for feature selection to provide the best
extracted features to avoid unnecessary ones and reduce calculations and retrieval
time. Meanwhile, the k-nearest neighbors algorithm is used to search for the relevant
images. Finally, we will apply our method on two different image data-sets which
contain images belonging to different domains to show the efficiency of selecting
the correct features using meta-heuristic algorithms.
1 Introduction
The image bases are becoming very wide and diversified. In order to well exploit
these large databases and extract the relevant images, content-based image retrieval
(CBIR) is the powerful method so far. This method which used the visual descriptors
of the image is able to find the closest images to a query image (QI) in wide image
data-set. All CBIR systems follow two steps as represented in Fig. 1. In the offline
phase, the CBIR system extract the visual features of the database images, then store
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 881
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_80
882 N. Machhour and M. Nasri
them in a new features database. In the next online phase, the CBIR system calculates
the similarity measures between the extracted features of all data-set images and
the QI ones to determine the similar images to the inputted QI. Therefore, color
descriptor which is used widely in CBIR systems provide good results. Yue et al. [1]
demonstrated that color features are introduced in pretty nearly all CBIR systems
due to their advantages, such as color histogram [2] and color moments such as the
mean, variance and third moments. Also, texture is an essential descriptor in CBIR
systems. The gray level co-occurrence matrix (GLCM) which give more details
about the arrangements of colors [3] can provide some important informations such
as: energy, entropy, contrast and homogeneity. Moreover, shape features are used to
describe particular regions in image. Thus, many techniques are developed such as:
edge detection, filters, Fourier transform or Hu moment.
There are many works developing CBIR systems which use a single image
descriptor such as color, as developed in the work of Sharma et al. [4] which
used the conventional color histogram (CCH) and Lei et al. [5] which applied a
quantization method on the HSV color space into 36 non-uniform bins, etc. These
studies give encouraging results but ignore other descriptors which can enhance the
retrieval precision. On another hand, several works merged a lot of features simulta-
neously, these methods can increase the search time and provide inefficient results by
using unnecessary descriptors. The work in hand used an efficient method based on
the feature selection using meta-heuristic algorithms applied to MPEG-7 standards
which increasingly become a general framework for content description. The aim
of this study is to enhance the precision rate and decrease the query time by main-
taining intelligently the relevant features from the MPEG-7 standards to avoid the
unnecessary descriptors which can influence the accuracy of the retrieved relevant
images and increase the response time by performing unnecessary calculations. For
this, the proposed CBIR system extracts the image features based on the MPEG-7
standards. Then applied the genetic algorithms (GAs) for feature selection to retain
the relevant descriptors. Finally, the k-nearest neighbors algorithm (k-NN) is applied
for image retrieval.
Image Retrieval Based on MPEG-7 Feature Selection … 883
2 Related Works
Multiple and various methods have been implemented in this domain. These studies
combined different visual features and used several search techniques. Alsmadi [6]
proposed a CBIR system using a technique based on the neutrosophic clustering algo-
rithm. They extract color descriptor based on the YCbCr using Canny edge histogram
(CEH) and discrete wavelet transform (DWT). The CEH is utilized to describe shape
characteristics and the GLCM to extract texture features. Machhour et al. [7], estab-
lished a method using the color string coding (CSC) technique combined to genetic
algorithm in the first work. Then, in an extended work, the authors in [8] improved
the color string coding technique by grouping each color series separately and devel-
oping a segmentation process based on the HSV color space. The GLCM is used
for extracting the texture characteristics. Meanwhile, they performed a comparison
between the simulated annealing (SA) and GAs in term of precision and response
time for image retrieval. In Ashraf et al. [9], the representation and image retrieval are
based on the artificial neural network (ANN) and the bandelet transform. ElAlami
[10] extracted the texture and color signature using the Gabor filter and the 3-D color
histogram. Then, they used the GAs to provide the optimal boundaries of the numer-
ical domains ranges with discrete values represented by nominal features. Finally,
they extracted the relevant features using 2 consecutive operations: the preliminary
and deeply reduction. In the aim of maximizing the precision and simplifying the
computation of the image search, the study in Lin et al. [11] is based on the feature
selection using the sequential forward selection technique which is developed on 3
characteristics: color co-occurrence matrix, pixels variance of scan pattern and color
histogram for k-mean.
The efficiency of a CBIR system is linked to several factors such as the features
extracted, retrieval method, response time, etc. As we know, many of these features
that are extracted from the image are irrelevant, which influence the quality of the
retrieved images and increase the calculation and response time. In the proposed
CBIR system as shown in Fig. 2, we implement a technique which is based on the
MPEG-7 standards to extract 6 descriptors representing the color, texture and shape
features of the image. Therefore, our technique is based on two phases. In the first
one which is done offline, we extract the image features which are: the dominant
color descriptor (DCD), color structure descriptor (CSD) and color layout descriptor
(CLD) for color features; texture edge histogram descriptor (EHD) and 2D-discrete
wavelet transform (2D-DWT) for texture features; and the curvature scale space
(CSS) for contour-shape descriptor. Then, we implement the genetic algorithms for
feature selection to avoid the unnecessary descriptors and retain the relevant ones.
884 N. Machhour and M. Nasri
In the second phase which is done online, we extract the query image features, and
apply the k-nearest neighbors algorithm (k-NN) to find the relevant images.
The color variance vi depicts the variation of the values of image elements assigned
to the class of the dominant color ci ; while s which represents the spatial coherency
shows the homogeneity of this dominant color ci .
Color Structure Descriptor (CSD). This descriptor is used to represent an image by
the color distribution [13, 14]. According to Messing et al. [14], this feature can be
defined as a generalization of the color histogram which can be represented by the
color structure histogram described by the formula below:
Where, M is the quantized color cm (M ∈ {256, 128, 64, 32}), and h(m) is the
bin value representing the number of the structuring elements having pixels with
color cm .
Hence, the CSD is based on the information of the spatial structure of the colors
and their frequency occurrence. Messing et al. [14] proved that the performance of
this descriptor can be enhanced based on the non-uniformly quantization histogram
amplitudes using the nonlinear HMMD color space.
Color Layout Descriptor (CLD). The discrete cosine transform (DCT) is developed
to extract this descriptor to describe the spatial distribution of colors. For this, Manju-
nath et al. [12] specified four steps which are: image partitioning technique, repre-
sentative color detection method, discrete cosine transform, and then, the nonlinear
quantization of the zigzag-scanned coefficients.
Texture Features Extraction
Edge Histogram Descriptor (EHD). This descriptor describes a range of 5 types of
edges in each part of image which is divided into n × n blocks. The 5 types of these
edges developed by Won et al. [15] and shown in Fig. 3 are: vertical (a), horizontal
(b), 45-degree diagonal (c), 135-degree diagonal (d), and non-directional edges (e).
Therefore, every local histogram contains 5 bins which correspond to one of 5
edge types. Won et al. [15] demonstrated that dividing the image into 16 blocks
provide a total of 80 histogram bins. Therefore, to extract these features, the image
is converted to gray level, then dividing this image into sub-blocks. Finally, these
descriptors are provided using the directional and non-directional edge features.
2D-Discrete Wavelet Transform (2D-DWT). This descriptor is very efficient in many
domains including image processing. This technique can decompose an image in
various frequency sub-bands. Also, 2D-DWT depicts the frequency and spatial range
and can be used efficiently for non-stationary signals. The technique is based on
dividing the image into rows and columns by applying the average intensity of the
image (which is illustrated by low pass) and high temporal resolution describing the
edges of image (high pass). Figure 4 illustrate an example of two-level decomposition
of the DWT. This method utilizes filters such as Haar, Daubechies, etc. In this study,
we used the low-pass (L) and the high-pass (H ) filters based on the Haar Wavelet
Descriptor as shown below:
L = [1, 1] (4)
H = [−1, 1] (5)
Shape Features Extraction. In this study, the shape features extraction is based on
the MPEG-7 standards. Thus, we utilize the curvature scale space (CSS) for contour-
shape descriptor. This latter is characterized by its robustness against noise, scale and
change in translation and orientation. Firstly, we select N -samples equidistant point
on the contour, next we choose randomly a point and follow the contour clockwise.
Then, we create two series (x) and (y) to group and store separately for each selected
point two coordinates (x and y coordinates) respectively. Consequently, we get an
arc-length parameterization depicted by the formula below where (u) is an arbitrary
parameter:
Where, μ represents the chromosome mean, xi the gene value and N the size of
the chromosome.
Natural Operations of the GAs. The roulette wheel technique is used in the selection
phase to generate randomly an array (Ri) containing values between 0 and 1. Then,
we calculate the selection probability (Pi) for each chromosome to provide the
cumulative probabilities (Qi). The comparison between (Ri) and (Qi) select the
chromosomes to yield the new generation.
In the next phase, the population is exposed to the crossover operation. The GA
chooses 2 chromosomes to exchange their genes between them from a crossing point
chosen randomly. The probability of the crossover operation (Pc) is 0.75, that means
that on average 75% of the population will undergo this operation.
The mutation operation is an essential process, because it makes random changes
in the population. Thus, all individuals in the population have the same chance to
be chosen and the mutation probability is 0.01. Then, the gene in a position chosen
888 N. Machhour and M. Nasri
In CBIR systems, many retrieval methods have been developed such as tradi-
tional techniques or new approaches using artificial intelligence techniques or meta-
heuristic algorithms. Machine learning methods have shown their effectiveness in
this type of problems. In this study, we develop the k-nearest neighbors algorithm
for image retrieval. This non-parametric and supervised machine learning algorithm
was introduced in 1967 by Thomas Cover et al. [18] and developed for regression
and classification problems. The k-NN algorithm is among the simplest methods in
machine learning; and it is used massively in several domains because of its ease of
implementation and use and low execution and calculation time. The steps followed
by the k-NN algorithm are:
– Load the database of selected features.
– Initialize the value of k.
– For each image in the data-set.
• Compute the distance value of the current image and the QI.
• Store this distance in the distance data-set.
– Sort the distance data-set.
– Select the first k images from the distance data-set.
In this work, we apply our method on two different image bases which are: Corel
and Caltech image data-sets. These large image bases contain hundreds of images
belonging to different domains; and allow us to evaluate and test the performance
of the feature selection method and compare the results obtained in this study with
other methods. For this, we calculate two measures: precision (P) and recall (R).
Image Retrieval Based on MPEG-7 Feature Selection … 889
Table 1 The average recall and precision values using Corel image data-set
Classes Africa Food Horses Elephant Flower Beach Building Nature Bus Dinosaur
(P) 0.76 0.76 0.84 0.84 0.88 0.76 0.76 0.76 0.88 0.96
(R) 0.19 0.19 0.21 0.21 0.22 0.19 0.19 0.19 0.22 0.24
Table 2 The average recall and precision sample using Caltech image data-set
Classes Minaret Tick Flamingo Bike Okapi Chandelier Ball Rhino Rooster Panda
(P) 0.92 0.84 0.80 0.80 0.76 0.72 0.96 0.72 0.80 0.80
(R) 0.30 0.42 0.29 0.32 0.48 0.16 0.37 0.30 0.40 0.52
The average precision and recall computed in this paper are: 0.82 and 0.205 for
Corel image data-set and 0.812 and 0.356 for Caltech image data-set respectively.
Table 1 and Table 2 represent the precision and recall rates computed for Corel and
Caltech data-sets respectively. Furthermore, Table 3 depicts a comparison of the
results found in this study and other CBIR systems using the Corel image data-set.
As evidenced by the results, the proposed method provides good recall and preci-
sion rates which shows the efficiency of feature selection technique based on GAs.
Furthermore, applying several times the GAs for feature selection, we notice that the
EHD descriptor is efficient in images including distinct objects and CSD descriptor
provide pertinent results in images containing principal colors. However, combining
the 3 best selected features in this study which are: CSD for color features, EHD
for texture descriptor and CSS for contour-shape descriptor, provide relevant results
for almost any type of images as shown in results above. Moreover, the use of the
feature selection reduces the number of descriptors which decreases enormously the
calculations, computations of image retrieval and execution time. Furthermore, the
k-NN algorithm is a relevant choice as image retrieval method because it provides
pertinent and satisfactory results in low execution and response time.
890 N. Machhour and M. Nasri
5 Conclusion
In the work in hand, we have developed an efficient CBIR system based on feature
selection technique. Once the QI is entered, the proposed system extracts 6 MPEG-
7 features for color, texture and shape signature. Meanwhile, these descriptors are
subject to natural operations of the genetic algorithm to retain the relevant features
and remove the unnecessary ones. Finally, the k-NN algorithm is applied for image
retrieval. The experiments are done using the Corel and Caltech image data-sets. As
illustrated in the previous section, the results obtained have demonstrated that the
GAs are powerful and intelligent method for selecting the correct and relevant image
features. Furthermore, removing the needless image characteristics has improved
enormously the retrieval time by avoiding additional calculations and increased the
accuracy of retrieved images. Also, the results found by the k-NN algorithm show
the efficiency of this machine learning technique. As depicted by the retrieval results,
especially the average recall and precision rates which are 0.205 and 0.82 respec-
tively, combining the MPEG-7 standards for image features extraction, GAs for
feature selection and the k-NN method for image search provide satisfactory results
which demonstrates and proves the efficiency of the proposed method which also
exceeds other proposed CBIR systems. In the prospective works, other techniques
and optimization methods will be utilized for feature selection; and combined to other
image descriptors while using different retrieval methods to increase the accuracy of
relevant images in content-based image retrieval system.
References
1. Yue J, Li Z, Liu L, Fu Z (2011) Content-based image retrieval using color and texture fused
features. Math Comput Model 54(3–4):1121–1127
2. Swain MJ, Ballard DH (1991) Color indexing. Int J Comput Vis 7(1):11–32
3. Shapiro LG, Stockman GC (2001) Computer vision. Prentice-Hall, Upper Saddle River
4. Sharma NS, Rawat PS, Singh JS (2011) Efficient CBIR using color histogram processing.
Signal Image Process 2(1):94
5. Lei Z, Fuzong L, Bo Z (1999) A CBIR method based on color-spatial feature. In: Proceedings
of IEEE. IEEE Region 10 conference. TENCON 99. Multimedia technology for Asia-Pacific
information infrastructure (Cat No 99CH37030), vol 1. IEEE, pp 166–169
6. Alsmadi MK (2020) Content-based image retrieval using color, shape and texture descriptors
and features. Arab J Sci Eng 45:3317–3330
7. Machhour N, M’Barek N (2020) Content based image retrieval based on color string coding
and genetic algorithm. In: 2020 1st international conference on innovative research in applied
science, engineering and technology (IRASET). IEEE, pp 1–5
8. Machhour N, Nasri M (2020) A novel content-based image retrieval based on a new approach
of color string coding and meta-heuristic algorithms. Int J Adv Trends Comput Sci Eng
9(1.5):128–137
9. Ashraf R, Bashir K, Irtaza A, Mahmood MT (2015) Content based image retrieval using
embedded neural networks with bandletized regions. Entropy. 17(6):3552–3580
10. ElAlami ME (2011) A novel image retrieval model based on the most relevant features. Knowl-
Based Syst 24(1):23–32
Image Retrieval Based on MPEG-7 Feature Selection … 891
11. Lin CH, Chen RT, Chan YK (2009) A smart content-based image retrieval system based on
color and texture feature. Image Vis Comput 27(6):658–665
12. Manjunath BS, Salembier P, Sikora T (2002) Introduction to MPEG-7: multimedia content
description interface. Wiley
13. Manjunath BS, Ohm J-R, Vasudevan VV, Yamada A (2001) Color and texture descriptors.
IEEE Trans circuits Syst Video Technol 11(6):703–715
14. Messing DS, Van Beek P, Errico JH (2001) The mpeg-7 colour structure descriptor: Image
description using colour and local spatial information. In: Proceedings 2001 international
conference on image processing (Cat No 01CH37205), vol 1. IEEE, pp 670–673
15. Won CS, Park DK, Park S (2002) Efficient use of MPEG-7 edge histogram descriptor. ETRI J
24(1):23–30
16. Mokhtarian F, Abbasi S, Kittler J (1996) Robust and efficient shape indexing through curvature
scale space. In: In British machine vision conference. Citeseer
17. Holland JH (1975) Adaptation in natural and artificial systems. University of Michigan Press,
Ann Arbor
18. Cover T, Hart P (1967) Nearest neighbor pattern classification. IEEE Trans Inf Theory 13(1):21–
27
19. Machhour N, Nasri M (2020) New color and texture features coding method combined to the
simulated annealing algorithm for content based image retrieval. In: 2020 fourth international
conference on intelligent computing in data sciences (ICDS). IEEE, pp 1–8
A Powerful and Efficient Method
of Image Segmentation Based
on Random Forest Algorithm
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 893
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_81
894 Z. Faska et al.
image that the measurements are carried out for the extraction of the discrimi-
nating parameters with a view to the classification or interpretation. Therefore, image
segmentation aims to find homogeneous regions and contours. A region is a set of
adjacent pixels, which are grouped into distinct and homogenous regions according
to pre-defined criteria (texture, color, etc.).We can summarize the segmentation task
as follows given an image; the objective of segmentation is to establish a compact
description that is representative of its information content, which is more useful
than all of its points. This involves extracting relevant visual indices (primitives),
sufficiently correlated with the entities that make up the scene from which the image
is taken. Moreover, image segmentation is a vast subject of study. In the last years,
several works dealing with this subject have been published in different fields as:
camera self-calibration [1–5], 3D reconstruction [6, 7] and Cryptography [8, 9],…
etc. We can distinguish the following four different segmentation techniques that
exist in the literature (Fig. 1):
In the contour segmentation approach, we consider that the primitives to be
extracted are the lines of contrasts separating regions of different and relatively
homogeneous gray levels, or else regions of different texture. In practice, it is a
question of recognizing the transition zones and of locating the border between
the regions as well as possible. In the contour-based segmentation approach, there
are three categories of methods are derivative, analytical, and deformable methods
[10, 11].
In the region approach a grouping, the process is repeated until all pixels in the
image are included in regions. This approach, therefore, aims to segment the image
based on the intrinsic properties of the regions. We can distinguish the following
methods: region growing, merge, the division into regions, and split/merge [12, 13].
The segmentation by region-contour cooperation arouses great interest. It
combines the two segmentation techniques (region-based, contour-based), takes the
Segmentation techniques
(a)
(b)
(c)
Fig. 2 Examples of images in the ADE20K dataset. The a row shows the sample images, b row
shows the annotation of objects, and the c row show the annotation of object parts
896 Z. Faska et al.
2 Proposed Approach
In this section, we will treat our proposed method that is used to combine two
complementary processes. The first process to make the segmentation of the contours
by combining the canny algorithm, Prewitt and Sobel’s filters. The second process
is a supervised classification segmentation by using the results of the first process,
which is Random forest. For this, we worked with the ADE20K Dataset [28, 29].
The flowchart of the method proposed is illustrated as follows (Fig. 3):
The Canny filter (or Canny detector) [10] is used in image processing for edge
detection. The algorithm was designed by John Canny in 1986 to be optimal according
to three clearly explained criteria:
• Good detection: the operator must detect as precisely as possible all the contours
of an image.
• Good localization: optimization of the precision with which the contour is
detected.
• Uniqueness: the contour should elicit a unique operator response.
These three criteria are expressed in the joint optimization of three functions for
defining the optimal linear filter for the detection of a stair-step under the assumption
of an independent additive noise signal.
Color image
Contour Extraction
Image segmented
If we consider that the filter has an impulse response ψ(x), these functionalities
are written:
∞
ψ(x)d x
Good detection : = 0 (1)
∞
−∞ ψ 2 (x)d x
ψ (0)
Good localization : = (2)
∞
−∞ ψ 2(x)d x
0
ψ (0) ψ(x)d x
Uniqueness : = k −∞ (3)
∞ ∞
−∞ ψ 2 (x)d x −∞ ψ 2 (x)d x
The solution that verifies these three criteria, proposed by Canny is as follows:
x x −x −x
ψ(x) = a1 e σ sin ωx + a2 e σ cos ωx + a3 e σ sin ωx + a4 e σ cos ωx (4)
Where the coefficients ai and ω are determined from the size of the filter. The
parameter is a parameter of great importance that we find in all other filters derived
from the approach of Canny. This scale parameter indicates how far short of two
parallel contours will be merged into one. Canny shows that the derivative of a
Gaussian is a good approximation of its filter.
Starting from other initial conditions than those of Canny, Deriche’s detector
meets exactly the same quality criteria as Canny’s, but which has an infinite impulse.
It could therefore be synthesized in a particularly efficient recursive manner [11].
The Deriche filter has a general expression of the form:
2
−α|X | 1 − e−α
ψ(x) = −C X e With c = (5)
e−α
Noise, interference between adjacent edges, and precision errors due to the discreet
character of the image definition grid are among the factors that can corrupt the
performance of edge detection in real images. To detect intensity transitions by
numerical differentiation, derivative methods are the most used. These local methods
scan the image with a mask defining the area of interest. At each position, an operator
is applied to detect significant transitions at the level of the chosen discontinuity
attribute.
898 Z. Faska et al.
Numerous contour extraction techniques exist in the literature. They can be clas-
sified into the following categories: Algorithms based on the gradient (or first-
order operators), Algorithms based on the Laplacian (or second-order operators),
Multi-scale algorithms, and Algorithms using sophisticated filters.
In the case of gradient-based algorithms (or first-order operators), the gradient of
an image gives the rates of change in gray level per unit distance in the directions of
the coordinate axes. It is defined as a vector characterized by its amplitude (related to
the amount of local variation in gray levels), and direction (orthogonal to the border,
which passes at the point considered.
We can therefore obtain a perfect knowledge of the image, which is calculated as
follows:
∂ I (x, y) ∂ I (x, y)
∇x = Et ∇ y = (6)
∂x ∂y
Thus, at each point of the image (x, y), we define two partial derivatives, along x
and y. The direction of the gradient vector maximizes the directional derivative, and
its norm is the value of this derivative.
The various existing variants, including the Roberts operator [30], provide
improvements to the discrete estimation of the gradient (derivative) and to the consid-
eration of the orientation of the contour. The Prewitt operator [31] and the very
popular Sobel operator [32] allow local estimation of the norm of the two-dimensional
spatial gradient of a gray-level image. They amplify regions of strong local variations
in intensity corresponding to the contours. These operators consist of a pair of 3 ×
3 convolutional masks. Applying each of the masks separately gives an estimate of
the horizontal and vertical components of the gradient by simple linear filtering to
moderate the effects of noise.
The Random Forest classifier is a powerful machine-learning tool for solving clas-
sification and regression problems. Uses adaptive (boosting) or random (bagging)
strategies. It consists of using a combination or an aggregation of a large number of
models while avoiding overfitting.
Bagging is an ensemble method introduced by Breiman in 1996 [33]. The word
Bagging is the contraction of the words Bootstrap and Aggregating. Bootstrap is
a principle of statistical resampling [34] traditionally used for the estimation of
quantities or statistical properties. The idea of bootstrap is to use multiple sets of
data sampled from all observed data using a random draw with replacement. Thus,
each elementary classifier of the set will be trained on one of the bootstrap samples
so that they are all trained on a different training set. Aggregating these classifiers
allows for a more efficient predictor.
A Powerful and Efficient Method of Image Segmentation ... 899
1 B
G(x) = G i (x) (7)
B i=1
|A ∩ B|
I oU = (9)
|A ∪ B|
The Intersection over Union (IoU) also called Jaccard index, is an evaluation
metric used to measure the accuracy of an object detector on a particular dataset. It
measures the degree of similarity between the target mask and our prediction output.
• The Dice similarity coefficient (DSC), also called the Sorensen - Dice index or
simply the Dice coefficient, is a statistical tool that measures the similarity between
two sets of data. This index has undoubtedly become the most widely used tool in
the validation of image segmentation algorithms, A, B ⊆ S e R n it is expressed
by:
900 Z. Faska et al.
2|A ∩ B|
DSC = (10)
|A| + |B|
Here, | A | is the number of elements of A. The index can vary from 0 (when A
and B are disjoint) to 1 (when A and B are equal).
• Entropy: was introduced by Shannon (1948), is a mesure of image information
content, which is interpreted as the average uncertainty of information source,
were the higher value of entropy means more detailed information [35].
L−1
Entr opy = − p f (i)log2 p f (i) (11)
i=0
With p f the ratio of the number of pixels to the value of the gray level over the
total number of pixels.
• Mutual information (MI): is a quantity that measures a relationship between the
original image and image segmented, were the higher value of MI means the
relationship between the images exists.
PAF (a, f )
MI = PAF (a, f )log (12)
af PA (a)PF ( f )
With PAF (a, f ) the joint histogram of the merged image F and the source image.
In order to show the quality of our approach on a set of images, which takes the
same size of the original image in all methods:
A Powerful and Efficient Method of Image Segmentation ... 901
– The living room image is 256 * 256 in TrueColor, the Seaside Image is 225 * 300
in TrueColor, and the Museum image is 256 * 256 in TrueColor.
The best result of segmentation is where the values of IoU and dice are close to
the value 1, while the values of MI and Entropy are high. The metric values of some
images presented in Fig. 4 are given in the following Tables 1, 2, and 3:
Tables 1, 2, and 3 presented the results of the different methods obtained from
the four measurements IoU, Dice similarity coefficient, MI, and Entropy, where the
value in Italic in each of the rows shows the best segmentation result obtained by the
three methods.
We have found that our proposed segmentation cooperative approach gives the
best segmentation result compared to other methods. From these tables, we notice
that the values of IoU and DSC obtained by our approach are better than the other
techniques, as well as the value of entropy. Therefore, from the comparison between
different methods, the proposed cooperative approach provides an improvement in
the image segmentation results compared to other methods.
902 Z. Faska et al.
4 Conclusion
References
16. Saber EA (1998) Integration of color, edge, shape, and texture features for 94 automatic region-
based image annotation and retrieval. J Electron Imaging 7(3):684–700
17. Wazarkar S, Keshavamurthy BN, Hussain A (2018) Region-based segmentation of social
images using soft KNN algorithm
18. Yang HY, Wang XY, Wang QY, Zhang XJ (2012) LS-SVMbased image segmentation usingcolor
and textureinformation. J Vis Commun Image Represent 23:1095–1112
19. Breiman L (2001) Random forests. Mach Learn 45:5–32
20. Dhanachandra N, Manglem K, ChanuImage YJ (2015) Segmentation using K-means clustering
algorithm and subtractive clustering algorithm
21. Khan SS, Ahmad A (2004) Cluster Centre Initialization Algorithm for K-means Cluster. Pattern
Recognit Lett 25:1293–1302
22. Khrissi L, El Akkad N, Satori H, Satori K (2020) Image Segmentation based on k-means and
genetic algorithms. In: Embedded Systems and Artificial Intelligence. Springer, Singapore, pp
489–497
23. Khrissi L, Akkad NE, Satori H, Satori K (2019) Color image segmentation based on hybridiza-
tion between Canny and k-means. In: 2019 7th Mediterranean congress of telecommunications
(CMT). IEEE, pp 1–4
24. Khrissi L, El Akkad N, Satori H, Satori K (2020) Simple and efficient clustering approach based
on cuckoo search algorithm. In: 2020 fourth international conference on intelligent computing
in data sciences (ICDS). IEEE, pp 1–6
25. Shotton J, Winn J, Rother C, Criminisi A (2006) TextonBoost: joint appearance, shape and
context modeling for multi-class object recognition and segmentation. In: Proceedings of
ECCV, pp 1–15
26. Viola P, Jones M (2001) Rapid object detection using a boosted cascade of simple features. In:
Proceedings of CVPR, pp 511–518
27. Winn J, Criminisi A (2006) Object class recognition at a glance. In: Proceedings of CVPR
28. Zhou B, Zhao H, Puig X, Fidler S, Barriuso A, Torralba A (2017) Scene parsing through
ADE20K dataset. In: Computer vision and pattern recognition (CVPR)
29. Roberts LG (1965) Machine Perception of Three Dimensional Solids. In: Tippet JT (ed) Optical
and electrooptical information processing. MIT Press, pp 159–197
30. Prewitt JMS (1970) Object enhancement and extraction. In: Lipkin BS, Rosenfeld A (eds)
Picture processing and psychopictorics. Academic Press
31. Sobel (1990) An isotropic image gradient operator. In: Freeman H (ed) Machine vision for
three-dimensional scenes, pp 376–379. Academic Press
32. Breiman L (1996) Bagging predictors. Mach Learn 24(2):123–140
33. Efron B, Tibshirani RJ (1993) An introduction to the bootstrap. Chapman & Hall, New York
34. Breiman L, Friedman JH, Olshen RA, Stone CJ (1984) Classification and regression trees.
Chapman and Hall, New York
35. Yang L, Guo BL, Ni W (2008) Multimodality medical image fusion based on multiscale
geometric analysis of contourlet transform. Neurocomputing 72:203–211
Digital Technologies in Battle Against
COVID-19
Towards Overlapped Objects
Recognition Using Capsule Network
1 Introduction
Artificial intelligence is making anything artificial not natural like machines have the
intelligence of humans, the ability to understand, think and learn. Machine learning
is a subset of artificial intelligence. Number of definitions of machine learning have
been presented based on everyone’s vision. The most powerful one is Arthur Samuel’s
definition in 1959: “machine learning: field of study that gives computers the ability
to learn without being explicitly programmed”. [1] Learning is a specific human
property, having machines with human properties will make our life much easier.
Machine learning has different types of algorithms. Two most important ones are:
supervised learning and unsupervised learning. Supervised learning is teaching the
computer how to do a task while unsupervised learning learns to do it but itself [2].
Machine learning serves many fields: Data Mining [3], Natural Language Processing
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 907
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_82
908 M. Mansouri et al.
(NLP) [4], and so on. One of the most powerful fields of AI is Computer Vision; it is
training machines to interpret the visual world like humans using images and videos.
We are interested in object recognition from images, an essential domain in
computer vision that recognizes objects from an image based on several tasks such as:
detection, segmentation and classification. Two types of approaches were developed
to recognize objects: conventional approaches and deep learning based approaches.
Conventional or traditional approaches are any machine learning methods or
approach before the appearance of Deep learning-based approaches. Deep learning
is a subset of machine learning inspired by the construction of the human brain. The
difference between these two concepts is that classical machine learning takes the
input data and parses it to determine its features, then trains the machine learning
with that data, eventually the machine will apply what it has learned to make a deci-
sion. While deep learning has come to ensure end-to-end learning, it means from the
input image to the output, it handles the both feature extraction and decision phases.
Object recognition has a major role in different life domains: security [5], space
exploration [6], medical diagnosis [7] and so on. Some works recognize food to help
people who suffer from obesity and diabetes to control their nutrition, by calculating
for example number of calories of the recognized food to control the taken quantity
by day [8–11] while others try to recognize animals and give all information you
need about them [12, 13]. There are researches in the traffic monitoring environment
to recognize traffic signs [14, 15]. And others to detect and recognize general objects
[16–18]. The common challenge in object recognition is the overlapped objects. All
those mentioned works have shown impressive results in recognizing simple objects
or multiple non-overlapped objects in an image, but it is very challenging to define
two or more objects that partially hide each other. The missing parts of objects will
make it unfamiliar for machines. There are methods that arrive to detect and classify
overlapped objects, those methods may ensure state of the art accuracy but they work
either on one type of object shape [19] or require specific equipment [20].
Among overlapped objects recognition methods, the new artificial neural network
Capsule network (CapsNet) has shown great results on overlapped digits dataset
multi-MNIST [21], a special version of Modified National Institute of Standards
and Technology (MNIST) dataset. CapsNet has improved high accuracy compared
to convolutional neural networks (CNN). It has come to solve some CNN limits,
but nothing is perfect. CapsNet requires more computational resources and it does
not show high performance when it recognizes objects from an image with complex
background. In this paper, we propose the use of capsule networks to extract features
and classify six categories of overlapped fruits. In order to overcome CapsNet limits
that consist in complex background, we propose extracting the foreground (object)
from the background using GrabCut [22] segmentation algorithm. Our contribution
will overcome the challenge of overlapped objects recognition and CapsNet limit.
This paper can be summarized as follows: Sect. 2 contains related works, Sect. 3
describes our proposed system, and conclusion and future works are presented in
Sect. 4.
Towards Overlapped Objects Recognition … 909
2 Related Works
In this section, several existing works have been chosen based on specific comparison
criteria, there are works on general object recognition and others on specific objects
such as food, traffic signs, animals, and so on. The existing works of our problematic:
“overlapped object recognition” are presented while some approaches of capsule
network have been studied as well. The summary of those works are collected and
presented in a synthesis table giving a brief explanation of their limitations according
to our problematic and contribution.
Yu et al. [8] propose a classifier to learn incrementally the user’s data and each
user has 300 new types of food. The system used convolutional neural network
(CNN) to extract features and the combination of nearest class mean (NCM) clas-
sifier and nearest neighbor (NN) classifier for classification, trained and tested with
FoodLog dataset achieving with that the state of the art accuracy. Zhou et al. [16]
discuss the detection of the main object in an image contains different objects,
the authors propose region convolutional neural network (RCNN) to recognize the
multi-objects in an image, based on four steps, search selective for detection, CNN
for feature extraction, support vector machine (SVM) for classification and Non-
maximal suppression(NMS) to remove redundant candidate region and find the best
position of object detection. To obtain the main object from all detected objects, they
put a system based on the ratio of object volume and the image, and the rarity of
all the objects to mark the main object. The experiment was trained and tested with
ImageNet2012 and Flickr8k datasets and achieves 90% accuracy. Hayat et al. [17]
constructed a deep learning framework based on CNN to extract feature and recognize
nine general objects, they worked with Caltech-101 dataset for training and testing,
their work contains also a comparison of the proposed system and variant approaches
of Bag of Words (BOW) based on SVM, the results show that the proposed method is
much better with an accuracy of 90.12%. Liu et al. [9] propose an improved CNN to
extract features and recognize food using two datasets UEC-256 and Food-101 and
achieving 76.3 and 77.4% consecutively. Huang et al. [12] present a deep learning
system for recognizing 27 kinds of birds from images. The authors used the Internet
of Birds (IoB) dataset and preprocessed the images by flipping, rotating, translating,
and noise removing, and then feature extraction and recognition have done with
CNN achieving 95.3% accuracy. Vennelakanti et al. [14] worked on traffic signs
detection and recognition. First, they preprocess their images by converting RGB
image to HSV image, then detect the signs using color and shape based detection,
and finally extract features and recognize signs using CNN, for training and testing,
they used Belgium and German data sets achieving 98% accuracy. Sadeq et al. [10]
present the problem of size estimation in food recognition. Authors propose a system
910 M. Mansouri et al.
that employs capturing solid and liquid food items with fixed posture and estimate
the volume with simple geometric information, the system detects the food item
with ellipse detector, and segment it with color based K-means clustering, Otsu
thresholding, and Canny edge Detector, then uses CNN for features extraction and
classification. The used dataset was FooDD reaching 96% for recognition and 4.46%
error for size estimation. Taheri et al. [13] propose a system that classifies 19 classes
of animals using score-level fusion: fusion of two classifiers, one based on features
extracted from CNN and second based on appearance and shape-based features, to
make a decision and give final results, authors used nearest neighbor (NN) classi-
fier, and preprocessed images with resizing, converting from RGB to grayscale and
histogram equalization. For training and testing they employed LHI-Animal-Faces
dataset and achieved 95.31% accuracy. Ege et al. [11] present a review of five size
estimation methods of food; we are going to focus just on the two last new methods.
The first method used the two backside cameras in the iPhone to obtain depth infor-
mation. Authors used the convolutional neural network U-NET for extraction and
segmentation, and UECFood-100 for training, the second method used grains of
boiled rice as a reference object to estimate the real size based on CNN and its own
dataset trained. The approach achieves 10% estimation error.
Yukitoh et al. [20] present the difficulty of overlapping objects, the upper object
changes its angle and the lower object is partially hidden which leads to difficulty
in object recognition. To solve that, authors used two RGB-D sensors to recog-
nize different kinds of overlapped bread from own dataset with simple background,
the opposite viewpoints of the sensors were concatenated by iterative closest point
(ICP) algorithm, to recognize the upper object, authors used posture adjustment
and the two sensors and achieve 96.9% accuracy, for lower object, they used the two
sensors according to the degree of hidden, they reach 84.4% accuracy for 50% hidden
object. Changhui et al. [19] focused on overlapped citrus recognition for automatic
harvesting; they used threshold segmentation in HSV space, gift-wrapping convex
hull algorithm and transform method for citrus fruit localization and reconstruction.
They used their own dataset, as a result, the positioning and reconstruction error
rate were 6.51 and 8.84% respectively. Zhou et al. [23] present a novel segmenta-
tion method for overlapped fish images, authors captured images in real time using
camera attached to a computer, then preprocessed image by extracting background
and filtered the image from noise, then segmented overlapped fish applying several
methods: determine overlapped area using a shape factor, detect corners using the
curvature scale space algorithm then extract skeleton Zhang-Suen algorithm to obtain
as a conclusion the intersection points and the segmented overlapped fish. Authors
obtained 90% in accuracy as results. Song et al. [24] propose an improved version of
non-maximum suppression algorithm (NMS) in order to detect overlapped objects.
Authors grouped detected boxes into clusters and selected the center cluster with the
Towards Overlapped Objects Recognition … 911
Kumar et al. [15] propose a system that detects and recognizes traffic signs with
Capsule Network; they preprocessed their input images by enhancing brightness and
contrast and augmented the dataset by rotation and flipping. They used Germany
traffic signs dataset and achieved 97.6% accuracy. Ahmed et al. [18] present a 3D
capsule network for 3D object recognition, they proposed four 3D capsule network
architecture, containing different numbers of convolution layers, a dynamic routing
algorithm which is used between a primary capsule layer and class layer, squash
function was added and a decoder for object reconstruction. They used ModelNet-
10 and ModelNet-40 dataset which contain general objects 3D models, achieving
91.37 and 89.66% accuracy respectively. In order to reduce computational time and
improve classification, Hoogi et al. [25] propose a novel Capsule Network architec-
ture that contains self-attention mechanism which select the salient regions, and
CapsNet analyze only those regions, the proposed system was applied on three
medical datasets, MNIST and SVHN and achieve 90 94 92% accuracy on liver,
brain, lung datasets respectively and 99.5% on MNIST and improve the baseline
CapsNet accuracy in SVM with 2.4%.
In our search of related works to our thesis, we chose several approaches, mentioned
in the previous section, based on different comparison criteria; as a result, we built
a table of synthesis as described in Sect. 4.5. The comparison criteria proposed are:
Object Type (OT). The type of the recognized object in the approach.
Detection Method (DM). Methods used for detection of the object from an image.
Segmentation Method (SM). Methods used for segmentation of the object from
an image.
Feature Extraction Method (FEM). The method used to extract features.
Classification Method (CM). The method used for classification.
Overlapped Objects (OvO). The existing or not of detecting and classifying
overlapped objects:
Method Used (MU). The method used to recognize the overlapped objects.
Background (B). Define if the background is complex or simple.
Datasets (D). The used datasets.
912 M. Mansouri et al.
3 Proposed Approach
Fig. 1 Proposed system architecture. The images are examples of the supposed input and output
of the system
and edges, the weight of edges is defined based on the probable definition of each
pixel, a graph-cut is run to find a new classification of the pixels, after several itera-
tions, the unknown pixels are classified either background or foreground. Users can
correct results by marking the missing parts for optimization [22].
Our main dataset is fruit-360 dataset [27], it is a vegetables and fruits dataset
containing 120 categories, 61488 images for training and 20622 for testing of
simple objects with simple background. We chose just six categories as a start:
apple, avocado, banana, orange, lemon and strawberry. In order to avoid overfit-
ting in our network where the model learns the training set too well so it fails in
predicting other data of the same object, we expanded the principal dataset with a
916 M. Mansouri et al.
maid dataset composed of collected images from the internet. Moreover, the second
dataset contains overlapped objects to enrich the training phase. We applied GrabCut
on the second dataset to extract only the foreground and put it on a simple back-
ground as shown in Fig. 2. In order to train CapsNet we combined images from the
two datasets, each training class contains 100 images, and each testing class contains
25 images. All images have been resized to 100 × 100.
In the training phase, we preprocess our dataset by extracting foreground from back-
ground using GrabCut, and then feed our CapsNet for training. The CapsNet model
contains three layers, and a decoder, first layer is a convolution layer with ReLu
activation function, second layer is a primary layer contain a squash and reshape
function, the feature maps from convolution layer output will be reshaped to vectors,
and then normalized their length between 0 and 1 with squash function. Third layer
is capsule layer, here we use routing between capsules algorithm, the capsules in
primary layer will predict the output of the capsule layer, if there is an agreement,
the capsule predicted will be activated, and the capsule weight will be augmented,
this stage will be repeated after a number of iterations. The decoder is based on three
fully connected layers with two ReLu activation functions and one sigmoid function
to reconstruct the input segmented object. Figure 3 represents the used architecture
References
1. Samuel A (1959) Some studies in machine learning using the game of checkers. IBM J Res
Dev 3(3):210–229
2. COURSERA Homepage https://www.coursera.org/learn/machine-learning/home/welcome
3. Zhao L, Kong W, Wang C (2020) Electricity corpus construction based on data mining and
machine learning algorithm. In: 2020 IEEE 5th Information Technology and Mechatronics
Engineering Conference (ITOEC), pp 1478–1481. IEEE
4. Lansley M, Kapetanakis S, Polatidis N (2020) SEADer++ v2: detecting social engineering
attacks using natural language processing and machine learning. In: 2020 International
Conference on Innovations in Intelligent Systems and Applications (INISTA), pp 1–6. IEEE
5. Liu B, Wu M, Tao M, Wang Q, He L, Shen G, Yan J (2020) Video content analysis for
compliance audit in finance and security industry. IEEE Access 8:117888–117899
6. Yang X, Nan X, Song B (2020) D2N4: a discriminative deep nearest neighbor neural network
for few-shot space target recognition. IEEE Trans Geosci Remote Sens 58(5):3667–3676
7. Szymkowski M, Saeed E, Omieljanowicz M, Omieljanowicz A, Saeed K, Mariak Z (2020) A
novelty approach to retina diagnosing using biometric techniques with SVM and clustering
algorithms. IEEE Access 8:125849–125862
8. Yu Q, Anzawa M, Amano S, Ogawa M, Aizawa K (2018) Food image recognition by person-
alized classifier. In: 25th International Conference on Image Processing (ICIP), pp 171–175.
IEEE, Greece (2018)
9. Liu C, Cao Y, Luo Y, Chen G, Vokkarane V, Ma Y (2016) Deepfood: deep learning-based food
image recognition for computer-aided dietary assessment. In: 14th International Conference
on Smart Homes and Health Telematics, pp 37–48. Springer, Cham (2016)
10. Sadeq N, Rahat FR, Rahman A, Ahamed SI, Hasan MK (2018) Smartphone-based calorie
estimation from food image using distance information. In 5th International Conference on
Networking, Systems and Security (NSysS), pp 1–8. IEEE, Bangladesh
918 M. Mansouri et al.
11. Ege T, Ando Y, Tanno R, Shimoda W, Yanai K (2019) Image-based estimation of real food size
for accurate food calorie estimation. In: Conference on Multimedia Information Processing
and Retrieval (MIPR), pp 274–279. IEEE, USA
12. Huang YP, Basanta H (2019) Bird image retrieval and recognition using a deep learning
platform. IEEE Access
13. Taheri S, Toygar Ö (2018) Animal classification using facial images with score-level fusion.
IET Comput Vision 12(5):679–685
14. Vennelakanti A, Shreya S, Rajendran R, Sarkar D, Muddegowda D, Hanagal P (2019) Traffic
sign detection and recognition using a CNN ensemble. In: International Conference on
Consumer Electronics (ICCE), pp 1–4. IEEE, USA
15. Kumar AD (2018) Novel deep learning model for traffic sign detection using capsule networks.
arXiv preprint arXiv:1805.04424
16. Yu L, Chen X, Zhou S (2018) Research of image main objects detection algorithm based on
deep learning. In: 3rd International Conference on Image, Vision and Computing (ICIVC) pp
70–75. IEEE, China
17. Hayat S, Kun S, Tengtao Z, Yu Y, Tu T, Du Y (2018) A deep learning framework using convo-
lutional neural network for multi-class object recognition. In: 3rd International Conference on
Image, Vision and Computing (ICIVC), pp 194–198. IEEE, China
18. Ahmad A, Kakillioglu B, Velipasalar S (2018) 3D capsule networks for object classification
from 3D model data. In: 5fnd Asilomar Conference on Signals, Systems, and Computers, pp
2225–2229. IEEE, USA
19. Changhui Y, Youcheng H, Lin H, Sa L, Yanping L (2017) Overlapped fruit recognition for
citrus harvesting robot in natural scenes. In: 2nd International Conference on Robotics and
Automation Engineering (ICRAE), pp 398–402. IEEE, China
20. Yukitoh M, Oka T, Morimoto M (2017) Recognition of overlapped objects using RGB-D sensor.
In: 6th International Conference on Informatics, Electronics and Vision & 2017 7th Interna-
tional Symposium in Computational Medical and Health Technology (ICIEV-ISCMHT), pp
1–4. IEEE, Japan
21. Sabour S, Frosst N, Hinton GE (2017) Dynamic routing between capsules. In: Advances in
Neural Information Processing Systems, pp 3856–3866
22. Rother C, Kolmogorov V, Blake A (2004) Grabcut: interactive foreground extraction using
iterated graph cuts. ACM Trans Graph (TOG) 23(3):309–314
23. Zhou C, Lin K, Xu D, Liu J, Zhang S, Sun C, Yang X (2019) Method for segmentation of
overlapping fish images in aquaculture. Int J Agric Biol Eng 12(6):135–142
24. Song Y, Li X, Gao L (2020) Improved non-maximum suppression for detecting overlapping
objects. In: Twelfth International Conference on Machine Vision (ICMV 2019), v ol. 11433,
p. 114330F). International Society for Optics and Photonics (2020).
25. Hoogi A, Wilcox B, Gupta Y, Rubin DL (2019) Self-Attention Capsule Networks for Image
Classification. arXiv preprint arXiv:1904.12483
26. Mukhometzianov R, Carrillo J (2018) CapsNet comparative performance evaluation for image
classification. arXiv preprint arXiv:1805.11195
27. Mureşan H, Oltean M (2018) Fruit recognition from images using deep learning. Acta
Universitatis Sapientiae, Informatica 10(1):26–42
The Spread of the Corona Virus Disease
(Covid-19) and the Launch of 5G
Technology in China: What Relationship
Abstract In this paper, we highlight the technology of the fifth generation of mobile
communications (5G), its impact on human health and its association with the emer-
gence of the Corona virus in China. We have found through this study, that this
technology uses frequencies below 6 GHz used also in other previous generations of
mobile networks and they do not lead to the emergence of dangerous diseases like
cancer or the corona virus, like what some people claim, and it turned out that this
virus comes from the pangolin, which transmitted the infection to humans in Wuhan,
China, in late 2019.
1 Introduction
The last 20 years have witnessed important changes in mobile phone networks with
a profound transformation from the second generation (2G) (which enables us to
communicate with voice), to the third generation (3G) (which was distinguished in
addition to voice communication data transmission), and then to the fourth generation
(4G) (Which has been known to provide and transfer large-scale communications)
[1]. In fact, from a phone initially designed to conduct a voice conversation between
two users without providing any service other than sending/receiving SMS, today’s
modern smartphone has become a true mobile data center that provides access to
several services and applications (browser, camera, internet and games). This gener-
alization of smartphone internet access and video call access are driving a higher
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 919
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_83
920 A. Moutaouakil et al.
data rate. That is why the mobile phone industry has chosen to develop the fifth-
generation network (5G) to provide a technical solution to the problem facing 4G
today [2, 3].
But as soon as China announced the launch of this technique, it received a series
of accusations, according to which it would have caused a group of diseases, such
as cancer, heart diseases, the most recent of which is the Corona virus, in reason
for using this technique for strong millimeter wave. For some, it has developed into
burning 5G towers and assaulting telecom worker [4].
Today, we are witnessing a wide spread of the Corona virus (Covid-19), which has
caused the death of one million of people around the world, young and old, and has
made countries confused and Sacrificing their economy by closing the borders in the
absence of a vaccine so far. The disease first appeared in November 2019 in Wuhan,
central China and It is the city where 5G was launched in the same month. The World
Health Organization (WHO) declared this epidemic a public health emergency of
international concern on January 30, 2020 [5]. The thing that made everyone wonder
about the source of this epidemic, how it spreads, how to prevent it and how to treat
it, and whether it has anything to do with the launch of 5G technology.
In this paper, we will study the fifth generation (5G) and how it works and what
damage it does, especially for human health, and then compare it with the symptoms
caused by the Corona virus and determine if there is a connection between them or
not.
2 5G Technology
Fig. 2 5G Spectrum
car, E health, Smart Cities, industrial process control, entertainment, and of course
the Internet of Things (IoT).
In order to provide and support the ever-increasing traffic and ever-increasing
demand for bandwidth, 5G will need to expand its frequency range beyond the bands
normally used in mobile communications. This includes both the frequency bands
below 6 GHz and the millimeter frequencies.
As shown in the following figure, the lower part of this frequency range, below
30 GHz, is preferred from the point of view of propagation properties. In addition,
it is important to understand that high frequencies, especially above 10 GHz, can
serve as a complement to the lower frequency bands, and will be primarily used to
provide additional system capacity and very wide band transmission for higher data
rates. Millimeter frequencies can be used for indoor applications to provide very
high-speed connections [7].
The two basic types are referred to as “sub-6 GHz” (which just uses frequencies
below 6 GHz) and “millimeter waves”, which are 24 GHz and higher (this name
comes from the fact that the lengths of individual radio waves can be measured in
millimeters). According to [8], the only types of 5G radio signals currently used in
5G networks in China are the less than 6 GHz variants].
To ensure, at millimeter frequencies or the frequencies sub-6 GHz, a high-speed
connection and the ubiquity of the urban network, 5G will use "massive MIMO"
(Multiple Input Multiple Output) technology using multi-antenna networks [9].
The design of a new antenna for 5G wireless devices is more demanding at MMW
frequencies where the flexibility of the system architecture, efficiency, reliability
and compatibility with MIMO systems are the main considerations for achieving
efficiency. High spectral frequency, as well as reduced multipath fading. The most
important requirements for 5G MIMO antenna are greater bandwidth to allow simul-
taneous operation of many system services, high gain to cope with high frequency
atmospheric attenuation and absorption.
The following figure shows examples of antennas used for 5G.
The spread of this kind of antennas elicited a cry from many scientists. Over 200
scientists and physicians demanded a 5G moratorium. These scientists highlight in
particular the risks of cancer, genetic damage and neurological disturbances due to
the use of this type of antenna which sends and receives millimeter waves, but the
922 A. Moutaouakil et al.
Fig. 3 Examples of 5G
antennas
truth is that the energy associated with these waves is too weak to cause cell damage
or break the weak energy bonds that maintain molecular clusters among living things.
As well that we have been using wireless signals less than 6 GHz around the
world for years, which China is currently using for fifth generation as we have said
in a large number of applications. For example, all existing 4G cellular networks use
signals in this range, as do WI-Fi (operating at 2.4 GHz and 5 GHz) and a home
microwave oven [10].
So, if these signals could cause a kind of metabolic change in animals or people,
as some say, then we all have had their effect long ago. It did not happen because due
to the signal strength used by all of these systems, there was no measurable negative
connection between these signals and our health, and there was no way that could
lead to something that would have created the corona virus.
3 Covid-19
In late December of 2019, some similar symptoms appeared in many people in the
Chinese city of Wuhan, which made doctors believe that it was a virus that infected
the city and research began to show that later we are really facing a global epidemic
that is spreading rapidly as the World Health Organization said.
Covid-19 is an infectious disease caused by a strain of the corona virus, called
SARS-CoV-22, contagious by human-to-human transmission via respiratory droplets
or by touching contaminated surfaces and then touching one’s face. The most
common symptoms are fever, cough and difficulty breathing, but they can cause
acute respiratory distress which can lead to death [11].
The Spread of the Corona Virus Disease (Covid-19) … 923
Covid-19 can be identified by this ring of small pointed proteins - called spicules
- which surrounds it. Once the viral particles enter the mouth, nose or eyes, they will
move to the back of the nasal passages, but also to the mucous membranes located
at the back of the throat. Thanks to their small spikes, the particles will cling to the
cell membranes of the respiratory tract, which produce a protein called ACE2. Once
hooked, they will release genetic material similar to DNA, called RNA, which will
enter the human cell. This genetic material will then "divert" cells from their usual
functions to give them a new role: to create and release millions of copies of the
virus.
The investigation into the animal origins of SARS-CoV 2, the etiological agent
of Covid-19, is still current. Bats, notably the species Rhinolophus affinis, and
the Pangolin are the only two animal species harboring corona viruses very close
to SARS-CoV-2. While the bat is the most likely source of the virus, although no
formal evidence exists, the role of Pangolin in the chain of transmission remains to
be determined [12].
A study, the publication of which has been accelerated in Nature, analyzes the
composition of the firm in several samples of pangolins seized during anti-poaching
operations. Despite its protected species status, the Pangolin is the mammal that
suffers the most from illegal trade. Its flesh is consumed in Asia and its scales are
used in Chinese medicine. Genetic analysis of the Pangolin samples identified six
strains of corona viruses, all of which belong to the same phylo-genetic group as
SARS-CoV-2, beta-corona viruses. At the genomic level, the genes of the six strains
of penguins are organized in the same way as those of SARS-CoV-2 [13, 14].
Scientists are therefore now certain of the appearance of Covid-19, a process
which has absolutely nothing to do with the installation of antennas in 5G phase
because the epidemic comes from an animal.
4 Conclusions
The world now knows a host of challenges, the most important of which is how to
tackle this epidemic, which is spreading wildly and killing people. As health concerns
become so frightening, it is easy to fall into the trap of thinking that everything is a
potential threat. Fortunately, the scientific development that the world knows answers
us about these fears, thanks to which we have learned that the source of this epidemic
is some animals, and there is no connection to launching 5G in its spread, and we
have also confirmed that this technology does not cause any diseases currently and
in particular they only use waves which frequency is below 6 GHz used by 4G and
other recent applications.
924 A. Moutaouakil et al.
References
1. Rappaport TS, Sun S, Mayzus R et al (2013) Millimeter wave mobile communications for 5G
cellular: It will work. IEEE Access 1:335–349
2. Khan F, Pi Z (2011) An introduction to millimeter wave mobile broadband systems. IEEE
Commun Mag 49:101–107
3. Haraz OM, Ali MMM, Alshebeili S, Sebak AR (2015) Design of a 28/38 GHz dual band
printed slot antenna for the future 5G Mobile communication networks. Paper presented at:
IEEE International Symposium on Antenna and Propagation & USNC/URSI National Radio
Science Meeting 10:1532–1533
4. Wang C-X, Haider F, Gao X et al (2015) Cellular architecture and key technologies for 5G
wireless communication networks. IEEE Commun Mag 52: 122–130
5. WHO: Statement on the second meeting of the International Health Regulations 2005
Emergency Committee regarding the outbreak of novel coronavirus (2019-nCoV) (2020)
6. Wen-Shyang C, Chun-Kun W, Kin-Lu W (1999) Square-ring microstrip antenna with a cross
strap for compact circular polarization operation. IEEE Trans Antennas Propag, vol 47, no 10,
pp 1566–1568
7. Cheng X, Tang C, Zhang Z (2019) Accurate channel estimation for millimeter-Wave MIMO
systems. IEEE Trans Veh Technol 68:5159–5163
8. Seker C, Güneser MT, Ozturk T (2018) A review of millimeter wave communication for 5G. In:
Proceedings of the 2nd International Symposium on Multidisciplinary Studies and Innovative
Technologies (ISMSIT), Ankara, Turkey, 19–21 October
9. Li B, Fei Z, Zhang Y (2019) UAV Communications for 5G and beyond: Recent advances and
future trends. IEEE Internet Things J 6:2241–2263
10. Huang J, Wang C-X, Feng R, Sun J, Zhang W, Yang Y (2017) Multi-frequency mmWave
massive MIMO channel measurements and characterization for 5G wireless communication
systems. IEEE J Sel Areas Commun 35:1591–1605
11. Li Q, Guan X, Wu P et al (2020) Early Transmission dynamics in Wuhan, China, of novel
corona virus infected pneumonia N Engl J Med
12. Zhu N, Zhang D, Wang W et al (2020) A novel coronavirus from patients with pneumonia in
China, 2019 N Engl J Med, 382 2020, pp 727–733
13. Lai C-C, Shih T-P, Ko W-C, Tang HJ, Hsueh PR (2020) Severe acute respiratory syndrome
coronavirus2 (SARS-CoV-2) and corona virus disease-2019 (COVID-19): the epidemic and
the challenges International Journal of Antimicrobial Agents, 105924
14. Parry J (2020) China coronavirus: cases surge as official admits human to human transmission.
British Medical Journal Publishing Group
Analysis of Covid-19 Impact on Gender
(Male and Female) in South Asian
Country Pakistan
Abstract The impact of Coronavirus (COVID-19) disease has been reported for
different countries. Males are at high risk of death due to COVID-19 as compared
to females. There is no data available for Pakistan, the South Asian region. We are
therefore undertaking this analysis to determine the gender effect on the outcomes of
COVID-19 in the South Asian nation of Pakistan. An online survey was performed
across Pakistan (Punjab, Sindh, Khyber Pakhtunkhwa, Balochistan) and national
figures from the Pakistani government. In total, N = 113 individuals were included
(92 males and 21 females). The descriptive statistics, association tests, and Pie chart
indicate that males are more impacted by COVID-19 as compared to females. The
doctor’s response indicates that 81.4% of males and 18.6% females are COVID-19
effective in Pakistan. Information from Pakistani national government statistics indi-
cates that the number of male cases are 78%. Therefore, from the entire studies, we
can infer that COVID-19 attacks more on males rather than females in the South Asian
nation Pakistan. However, more quantitative studies are recommended to measure
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 925
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_84
926 S. Malik et al.
the extent to which sex would result in COVID-19 outcomes among Pakistan’s South
Asian countries.
1 Introduction
Table 1 COVID-19 case all over the South Asian Country Pakistan as per July 17, 2020 [11]
Sr. No Province Confirm cases Active cases Deaths Recoveries
1 Sindh 1099,000 34,070 1,922 74,076
2 Punjab 88,539 22,149 2,059 64,815
3 KPK 31,217 7,749 1,124 22,613
4 Balochistan 11,322 3,096 128 8,161
5 Gilgit Baltistan 1,775 340 391 396
6 AJK 1,808 685 46 1,077
928 S. Malik et al.
Fig. 2 Sex-disaggregated data of confirmed COVID-19 across the whole world [14]
the belief that there is a strong biological pattern, responsible for the higher rate of
deaths in humans, irrespective of country-specific demographics and approaches to
science (Fig. 3) [12, 16–18]. The COVID-19 ratio of males and females in the table
below indicates the gender equality trend.
In Fig. 4 hospitalizations, Intensive Care Unit (ICU) deaths, and CFR are listed
in multiple nations. However exact numbers are given. Throughout all indicators,
sexually disaggregated data was not accessible [18] (Fig. 5).
Fig. 4 Gender difference of COVID-19 in Europe (Italy, Germany Switzerland, Spain, France) and
East Asia (China)
Fig. 5 The male vs female data varies significantly by East Asian countries (China & North Korea),
European countries (Italy, Spain, Germany, Sweden), Oceania countries (Australia), Middle East
country (Iran), and North American country (Canada) [14]
930 S. Malik et al.
During past coronavirus epidemics, the outcome of the outbreak was worse
for men than for women: men had worse outcomes with Serious Acute Respira-
tory Syndrome (SARS) and a higher risk of dying from Middle East Respiratory
Syndrome (MERS) [15]. Pakistan is unique in this collection of results, there are
concerns around the accuracy of COVID-19 coverage of females. Overall, though
more studies, the finding that COVID-19 is just like widespread for males and females
all over the world would need to be confirmed. People diagnosed with COVID-19
are more likely to die in all states that record rates of death by gender.
The data in Fig. 6 showed that for every 10 females worldwide there are more cases,
death, hospitalization, and ICU admissions of males. Male had worse results than
females in coronavirus or serious respiratory syndromes (SARS) attacks. Statistics
from several countries throughout the globe propose that males would be at high risk
of death due to COVID-19. East Asian countries (China & North Korea), European
countries (Italy, Spain, Germany, Sweden, Switzerland, and France), Oceania coun-
tries (Australia), Middle East country (Iran), and North American country (Canada)
has a higher number of male cases than female attacked by COVID-19 [18, 19]. But
there is no data available for the South Asian country Pakistan. Hence, we conduct
this study to evaluate the sex gender impact on COVID-19 results in the South Asian
country Pakistan.
2 Methodology
We have collected data from different sources for this analysis. Firstly, we obtained
data on 23rd march from doctor’s doing jobs in hospitals (n = 113) all over Pakistan
(Punjab, Sindh, Khyber Pakhtunkhwa, Balochistan). All the doctors associated with
private and government hospitals are included in the study. We excluded nurses,
healthcare workers, and the doctors, which are doing house jobs. We obtained data
Analysis of Covid-19 Impact on Gender … 931
from the doctors with the help of a Google form-based questionnaire. In this data
we apply descriptive statistics first and then association tests to achieve the goal of
gender and sex influence on COVID-19 outcomes in Pakistan. Secondly, the data
on sex-specific percentages for Pakistan was collected directly from the national
government statistics of 25 April 2020 and African population of health center 25
January, 2021, which are added to the aggregated data issued by the Government of
Pakistan on 27 May 2020.
3 Results
In Table 2 expresses the results, which are obtained from our first source from the
doctors. The Table 2 describes the distribution of frequencies for which COVID-19
is more influenced by gender. Here, 81.4% of men respondents were impacted and
COVID-19 impacted 18.6% of women’s respondents in the South Asian country
Pakistan. The variable’s mean and standard deviation is (1.1858) and (.39071),
respectively.
The below pie chart diagram explains our results. As shown in Fig. 6 shows that
more males are affected by COVID-19 than females in the South Asian country
Pakistan (Fig. 7).
Secondly, the data from the national government statistics of 25 April 2020 and
data issued by the Government of Pakistan on 27 May 2020 was analyzed. Our
analyses indicate that approximately 70% of all COVID-19 cases in the South Asian
country of Pakistan are male, with very close statistics for sex-specific hospital admis-
sion, mortality, and post-infection recovery as shown in Table 2. Although the number
of male cases in Pakistan are 78%. From across South Asian nation Pakistan, with
sex-specific statistics on the number of all COVID-19 related deaths, the rate of
male deaths ranged from 65–76%. In the South Asian nation of Pakistan, the fatality
pattern indicates that the male-to-female event fatality ratio is 0.9 (1.9% for males
vs. 2.1% for females). In the below table Case Fatality represents the percentage of
positive cases that die afterward (Figs. 8, 9, 10 and 11).
Fig. 7 Pie chart of gender affected by COVID-19 in South Asian country Pakistan
Fig. 8 COVID-19 cases and deaths rates per 100,000 in men/women in Pakistan
4 Discussion
Researcher [31] mentioned in his research that worldwide, medical statistics warn
us that males suffer more from diseases of the respiratory system than females
(25%), particularly those caused by acute viral infections. Females are less vulner-
able to infectious diseases than Men, relying on distinct innate immunity, gonad
development of steroid hormones, and sex chromosome related factors.
Researcher [32] discusses in his study that there was increased predisposition
and susceptibility to heart disease in clinical trials where testosterone levels were
Analysis of Covid-19 Impact on Gender … 933
Fig. 9 COVID-19 Cases by age and sex (rates per 100,000) in Pakistan
high, which is more common in men who are more prone to COVID-19. Those
devoid of COVID-19 cardiac disease have a better prognosis associated with patients
with heart failure and hypertension [32]. Discussion in his paper that the sex and
gender inequalities found in COVID-19 prevalence highlight the need to consider
the effect of gender and sex on the occurrence of the disease and death, and to adapt
gender and sex care to fit. Experiences from previous diseases and pandemics have
strongly illustrated the value of integrating a gender and sex study into health policy
preparedness and response efforts.
Nevertheless, the gendered effects of illness rise in, epidemics or diseases have
not yet been discussed by legislation and public health initiatives. Such countries do
not disaggregate data by ethnicity and age as other countries do. Also, policymakers
will disaggregate and evaluate data for gender and age gaps in all countries. In fact,
934 S. Malik et al.
as prophylactic and clinical trials proceed. Many studies identify the autoimmunity
effect of sex hormones.
In addition, data from many studies also shows that male COVID-19 patients and
other acute viral infections have greater vulnerability to mortality than females. In
men compared with women, it is considered to be more prevalent. Researchers [21]
explain in his study that there are many reasons behind this variance, like behavioral,
genetic, hormonal, as well as immunological variables. The mortality rates for men
are two to three times higher than for women, but tend to fluctuate in feature of popu-
lation area and demography. Social variables linked to ethnicity, immune differences,
hormonal variations and lifestyle behaviors including smoking and drunkenness will
play a role. Researcher [22] illustrates in his research that male populations can in any
way be vulnerable to the COVID-19 pandemic as a result of genetic polymorphisms
identified by gender.
Researcher [23] stated in his paper that beyond the biological danger the disease’s
epidemiological evidence of transmission indicates that the area is not similar and
that its effect on multiple populations is markedly interdependent. Researcher [24]
justify the occurrence of COVID-19 more on males, researcher stated based on its
socio-economic status the risk of males contracting and dying from SARS-CoV-2 is
significantly increased. In England and Wales, males throughout the deprived areas
have a mortality rate of 77.6 per 100,000 (44.3 per 100,00 for females) rising to 172
per 100,000 in its most deprived areas (97 per 100,000 for females).
Data shared by a globally chosen company Global Health 50/50 encouraging
balance between men and women in health, found a higher share of deaths in men
than in women in virtually every region. The details in the Italy Integrated Monitoring
Newsletter (April 23, 2020 update reveals that men’s deaths are about twice as many
as women’s deaths (17.1 vs. 9.3%). Researcher [25] found similar results in Greece,
the Netherlands, Denmark, Belgium, Spain, China and the Philippines. A sample
of 4,880 respiratory patients or near contact with Covid-19 patients in a hospital in
Analysis of Covid-19 Impact on Gender … 935
Wuhan showed that positive rates of SARS-CoV-2 in men and older populations were
substantially higher (>70 years) but only age is defined as a risk factor (>70 years).
Likewise, a recent retrospective observational analysis of researcher [26] showed
that 67% of critically ill patients with SARS-CoV-2 were men. Furthermore, a study
of researcher [27] results for 1.099 Covid-19 patients found 58.1% to be males of 173
extreme incidents, 57.8% have affected this population. Furthermore, reported infor-
mation from survival research [28] found that male deaths are considerably higher and
symptoms greater than female deaths. Finally, the researcher [29] recently recorded
a 1.7-fold higher case fatality rate for males than females (P < 00001). However,
researchers [30] recommend a walk-through gate for mitigating the COVID-19 effect.
5 Conclusion
We conduct this study to evaluate the sex gender impact on COVID-19 results in the
South Asian country Pakistan. Mixed methodology was used in this study, primary
data from doctors doing jobs all over Pakistan (Punjab, Sindh, Khyber Pakhtunkhwa,
Balochistan) and secondary data from national Pakistani government statistics. The
doctor’s response shows that 81.4% of males and 18.6 females are affected by
COVID-19 in Pakistan. Descriptive analysis between gender and COVID-19 results
in a mean of 1.1858 and a standard deviation of .39071. The data from national
Pakistani government statistics show that the number of male cases in Pakistan
is 78%, the rate of COVID-19 related male deaths ranged from 65–76%. From
the whole analysis, we can conclude that COVID-19 attacks more on males rather
than females in the South Asian country Pakistan. East Asian countries (China &
North Korea), European countries (Italy, Spain, Germany, Sweden, Switzerland, and
France), Oceania countries (Australia), Middle East country (Iran), South Asian
country (Pakistan) and North American country (Canada) has a higher number of
male cases than female attacked by COVID-19. However, we recommend further
comprehensive studies to measure the degree to which sex will lead to COVID-19
results among South Asian country Pakistan.
Limitations of the Study
The key limitation of this study applies to the comparatively low number of
samples carried out in the field, indicating that the overall number of tests and figures
can be significantly higher than what we report. However, we do not expect the central
point of this study to change considerably.
Acknowledgment The authors were very thankful to Dr. Imran Hussain, Dr. Wajid Yaqoob Bajwa,
Dr. Sheerbaz, Dr. Ali Hassan Shah, and Dr. Saqib Ishaq for providing data for our research.
Conflict of Interest Statements Author Saba Malik (Corresponding author), Mubbra Habib,
Mehmood Ahmed Husnain Hashmi, Muhammad Tariq saeed, Anwaar ul huda, Syeda Anam zahra,
Muhammad Nisar, Muhammad Sajid, Shahbaz Ahmad Shahzad, Saddam Hussain, and Muhammad
Sakandar Majid declares that they have no conflict of interest.
936 S. Malik et al.
References
1. Guan WJ, Ni ZY, Hu Y, Liang WH, Ou CQ, He JX, et al (2020) Clinical characteristics of
coronavirus disease 2019 in China. N Engl J Med. https://doi.org/10.1056/NEJMoa2002032
2. Who director-generals opening remarks at the media briefing on COVID-19 - 11 March 2020.
World Health Organisation (2020). https://www.who.int/dg/speeches/detail/who-director-
generals-opening-remarks-at-the-media-briefing-on-COVID-19%2D%2D-11-march-2020.
Accessed 22 Mar 2020
3. Coronarvirus Resource Center, Johns Hopkins University Medicine. 2020. Johns Hopkins
University Medicine. Medicine JHU. https://coronavirus.jhu.edu/map.html. Accessed 22 Mar
2020
4. Wu C, Chen X, Cai Y, Xia J, Zhou X, Xu S, et al (2020) Risk factors associated with acute
respiratory distress syndrome and death in patients with coronavirus disease 2019 pneumonia
in Wuhan, China. JAMA Intern Med. https://doi.org/10.1001/jamainternmed.2020.0994
5. Wang D, Hu B, Hu C, Zhu F, Liu X, Zhang J, et al (2020) Clinical characteristics of 138
hospitalized patients with 2019 novel coronavirus-infected pneumonia in Wuhan, China. Jama.
https://doi.org/10.1001/jama.2020.1585
6. Grasselli G, Pesenti A, Cecconi M (2020) Critical Care Utilization for the COVID-19 Outbreak
in Lombardy, Italy: Early Experience and Forecast During an Emergency Response. Jama.
https://doi.org/10.1001/jama.2020.4031
7. Zhao S, Cao P, Chong MK, Gao D, Lou Y, Ran J, et al (2020) The time-varying serial interval of
the coronavirus disease (COVID-19) and its gender-specific difference: a data-driven analysis
using public surveillance data in Hong Kong and Shenzhen, China from January 10 to February
15, 2020. Infect Control Hosp Epidemiol.https://doi.org/10.1017/ice.2020.64:1-8
8. Mo P, Xing Y, Xiao Y, Deng L, Zhao Q, Wang H et al (2020) Clinical characteristics of refractory
COVID-19 pneumonia in Wuhan. China Clin Infect Dis. https://doi.org/10.1093/cid/ciaa270
9. Conti P, Younes A (2020) Coronavirus Cov-19/Sars-Cov-2 affects women less than men:
clinical response to viral infection. J Biol Regul Homeost Agents 34(2)
10. Cai H (2020) Sex difference and smoking predisposition in patients with COVID-19. Lancet
Respir Med. https://doi.org/10.1016/s2213-2600(20)30117-x
11. World Health Organization. WHO Coronavirus Disease (COVID-19) Dashboard, 17 July 2020
https://covid19.who.int/. Accessed 17 July 2020
12. The epidemiological characteristics of an outbreak of 2019 novel coronavirus diseases (COVID-
19)—China, 2020. 2020. The Novel Coronavirus Pneumonia Emergency Response Epidemi-
ology Team, China CDC Weekly. https://www.ne.jp/asahi/kishimoto/clinic/cash/COVID19.
pdf. Accessed 03 Apr 2020
13. Globalhealth. (2021). The COVID-19 Sex-Disaggregated Data Tracker | Global Health 50/50.
https://globalhealth5050.org/the-sex-gender-and-covid-19-project/the-data-tracker/. Accessed
18 April 2021
14. Karlberg J, Chong DS, Lai WY (2004) Do men have a higher case fatality rate of severe acute
respiratory syndrome than women do? Am J Epidemiol 159(3):229–231. https://doi.org/10.
1093/aje/kwh056
15. Chen X, Chughtai AA, Dyda A, MacIntyre CR (2017) Comparative epidemiology of Middle
East respiratory syndrome coronavirus (MERS-CoV) in Saudi Arabia and South Korea. Emerg
Microbes Infect 6(6):e51. https://doi.org/10.1038/emi.2017.40
16. Epidemia COVID-19 (2020) Aggiornamento Nazionale. Dipartimento Malattie Infettive e
Servizio di Informatica, Istituto Superiore di Sanità. https://www.epicentro.iss.it/coronavirus/
bollettino/Bollettino-sorveglianza-integrata-COVID19_30-marzo-2020.pdf. Accessed 30 Mar
2020
17. Coronavirus disease 2019 (COVID-19) daily situation report of the Robert Koch Institute 2020.
Robert Koch Institute. Institute RK. https://www.rki.de/DE/Content/InfAZ/N/Neuartiges_C
oronavirus/Situationsberichte/2020-04-01-en.pdf?__blob=publicationFile. Accessed 03 Apr
2020
Analysis of Covid-19 Impact on Gender … 937
1 Introduction
COVID-19 has been an epidemic disease caused by the foremost recently discov-
ered Severe Acute Respiratory Syndrome CoronaVirus2 (SARS-CoV2) [1]. There
are different diseases find by coronaviruses. Including a Severe Acute Respiratory
Syndrome (SARS), discovered from musk cats to humans in 2003, the Middle East
Respiratory Syndrome (MERS) discovered from dromedary to humans in 2012, and
COVID-19 was caused by bats [2]. Coronavirus first unveiled in Wuhan, December
2019. COVID-19 is presently an epidemic affecting many nations globally. With the
information of the World Health Organization (WHO) approximately in ten months,
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 939
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_85
940 V. Rohini et al.
there are more than 35.4 million people infected across 214 nations and territo-
ries. Coronavirus caused more than 1.04 million deaths, 24.6 million individuals
have recovered. The number of individuals infected by this Coronavirus disease is
increasing, and it turns to be increasingly hard to end the diagnostic process within
the constrained time. Currently, many efforts and more attention paid to identifying
the COVID-19 affected patients. The COVID-19 most common signs and symptoms
are dry coughs, fever, tiredness and less common signs and symptoms are aches and
pains, a headache, sore throat, diarrhea, conjunctivitis, lack of taste or smell, a rash
on the skin or discoloration of palms or toes [3]. Signs and symptoms are difficulty
respiration, chest pain, and lack of speech or movement. This virus is unfolding
among people at some point of contact, commonly through tiny droplets through
coughing, sneezing and talking. This contamination can also additionally attain to
pneumonia. Lungs was effected by the pneumonia infection and also in the human
body lungs was mostly effected by the coronavirus. So that CT scan and Chest-x-ray
images are used to diagnose the pneumonia.
Therefore, it requires expertise and necessitates for different algorithms to analyze.
Consistent with this context, Machine Learning algorithms have obtained better
performance in detection of pneumonia, and provides accurate detection rate. By
using machine learning algorithms CT scan method is used to diagnose the chest
pneumonia [4]. But it is the time taken to process and cost-effective. And CT scanner
may not available in many regions as well. For that purpose, the contribution of our
paper used chest-x-ray images because chest-x-ray images are widely available and
take less time than the CT scan [5, 7]. So that, our paper follows penalized tuned
versions of Convolution Neural Network-based models InceptionV3, Resnet50.
A deep convolutional neural network used to examine the visual images and also for
classifying images. Convolutional Neural Networks (CNNs) is the used weights for
each pixel are fully connected. The layers in the convolutional neural network are
fully connected, pooling and Convolutional layers. These layers are used to perform
a different tasks on the input data. When deep transfer learning [6] techniques applied
to medical data, meaningful results are come out and these techniques [7] are used
for classification, segmentation and infection detection of medical data.
2 Related Work
Here some related research work for COVID-19 disease. S. Sai Thejeshwar et al. [8]
presented a KE Sieve Neural Network using the transfer learning for features which
extracted and weights on the VGG-19 pre-trained CNN model. This algorithm is not
capable of some specified dataset.
Deep Residual Convolutional Neural Network … 941
Al-Bawi, Ali, et al. [9] the classical VGG neural network performed with the
convolutional COVID block, and it obtained 95.51% accuracy for covid-19, normal,
pneumonia chest x-ray images. It was time taking process, and training deep neural
network was difficult.
Fang Y. et al. [10] proposed an RT-PCR technique used to identify the viral
RNA and serology testing will identify the visible representation of antibodies. In
this study when observing the patient samples, there is a difference in the detection
rate for initial CT (50/51 [98%]) patients greater than first RT-PCR (36/51 [71%])
patients. Collected annotated data is very expensive or difficult validation CT datasets
collected in one system developed to quantify infections only.
Zhang, Jianpeng, et al. [11] Rapid and accurate detection of the virus used
pervasive prevention and large-scale screening. Without fine-tuning, anomaly three
modules achieved an accuracy of 83.61%. It is cost-effective.
Shan F, Gao Y, Wang J, et al. [12] Deep learning-based method used for segmen-
tation and quantification of the respiratory system using chest CT scans [4]. This
system used a “VB-Net” neural network obtained dice similarity coefficients 91.6%
± 10.0% and with accuracy 86.7% on CT images. It is less sensitive to detect virus
from the throat.
Mohamed Loey et al. [13] GAN (Generate Adversarial Network) is for prepro-
cessing, deep transfer model for training, testing and validation. GAN has Image
Generation network and discriminator network. The Image Generation network
generates false data. Discriminator network describes the difference between real
and fake data. The algorithm takes some set of transfer models. The considered
dataset was 30 times larger than the original dataset, so it creates the fake dataset
images.
3 Proposed Methodology
In this method, the deep convolutional neural network (DCNN) based pre-trained
models have used for the categorization of chest-x-ray images. By using the Deep
Transfer learning method, it controls inadequate data and training time. By using
the pre-trained models of the CNN i.e., ResNet50 [14] and InceptionV3 [15] the
coronavirus was detected from chest-x-ray images.
ResNet50 is the best technique [16] of CNN because it uses shortcuts and also
saves the time of training. It intercepts the distortion of a system which is deeper
and complex. In the ResNet-50 model, bottleneck blocks used for faster training
[17]. ResNet50 contain 50-layer deep neural network and it has trained on the Image
Net database [18]. The database has contained 14 million images created for image
recognition competitions which are belonging to more than 20000 categories [19].
InceptionV3 is CNN pre-trained model. It consists of several convolutional steps
and maximum pooling steps. A Fully connected convolution network has connected
at the final stage. The InceptionV3 pre-trained model used to extract the features of
images [20] (Fig. 1).
942 V. Rohini et al.
Chest-x-ray images
Cross-Validation
Training dataset
Validation dataset
Estimated
Probability
The input dataset has taken from open source Kaggle dataset repository [21]. The
repository contains covid-19 and regular chest-x-ray images. It is an unbalanced
dataset. Here the output is a probability of accuracy (Figs. 2 and 3).
The introduced method using python3 programming language with Ubuntu 16.04
operating system. The considered dataset contains chest x-ray images. The 30 epochs
carried out to over fitting pre-trained models. The dataset differentiated into 2 parts
80% training and 20% testing for performance evaluation. The 5-fold cross-validation
the method is chosen k value from 1 to 5 (Table 1).
1
5
performance = performancek
5 k=1
In this proposed work, data is trained and tested on the chest-x-ray images. Training
accuracy and loss values of pre-trained ResNet50, InceptionV3 for fold-3 shown
in Figs. 4 and 5. To avoid the over-fitting 30epochs are taken for the training.
When compared to InceptionV3, Resnet50 obtained the highest training accuracy.
The ResNet50 gives a fast training process than InceptionV3. The loss values are
decreased in ResNet50 and InceptionV3 at the training stage. ResNet50 pre-trained
model reduces the training loss values quickly than the InceptionV3.
The other detailed way of given test data shown in Table 2. The highest accuracy of
98.2%. In training and testing, ResNet50 provides superiority over the InceptionV3.
Here when observe the given table confusion matrix results are described from the
pre-trained models fold-1 to 5.
5 Conclusion
For the detection of coronavirus, the proposed method deep convolutional neural
network based approach using chest-x-ray images has obtained a good result on the
considered dataset automatically. ResNet-50 is the one of best algorithm in deep
learning having a good result. The performance of ResNet50 pre-trained model
attains with the highest accuracy 98.2%. It is an accurate, rapid and effective clinical
support system for covid-19 detection. The training dataset is limited for proposed
work. Further, will try to work on different CNN models can be tested by using the
more number of Covid-19 and chest-x-ray images in the dataset.
References
1. Yan L, Zhang HT, Xiao Y, Wang M, Guo Y, Sun C, Tang X, Jing L, Li S, Zhang M, Xiao Y, Cao
H, Chen Y, Ren T, Jin J, Wang F, Xiao Y, Huang S, Tan X, Huang N, Jiao B, Zhang Y, Luo A,
Cao Z, Xu H, Yuan Y (2020) Prediction of criticality in patients with severe covid-19 infection
using three clinical features: a machine learning-based prognostic model with clinical data in
Wuhan. MedRxiv 2020.02.27.20028027
2. Costa S, Gomes, V, Moreli ML, Saivish, MV (2020) The emergence of SARS, MERS and
novel SARS-2 coronaviruses in the 21st century
946 V. Rohini et al.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 947
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_86
948 O. El Majdoubi et al.
1 Introduction
The theory of computing complexity reveals that several problems are complex and
proven to be not efficiently processed by sequential or parallel deterministic algo-
rithm [5]. The interest of the researchers surrounds the proposal of methods to solve
complex problems in order to reach optimal or sub-optimal solutions within a reason-
able run time. The exact methods are dedicated for small problems. Thus, the approx-
imate methods including heuristics and metaheuristics achieve solutions within a
reasonable execution time [6]. Commonly, there is no efficient method to attain opti-
mality in a quick polynomial run time for resolving NP-Hard problems, including
graphic coloring, Knapsack problem, integer programming, TSP and its variants. It is
imperative to speed up the execution time of proposed resolving methods, in partic-
ular for large scale problems [7]. The embedding of parallel processing techniques
is recommended and possibly depending on parallel architecture of the machine,
for resolving expensive algorithms and providing feasible solutions [8]. TSP is a
classical optimization problem, applied to many area contexts of computing, logis-
tics and industrial engineering. For its various applications, the researchers solved
this problem through branch and bound method, linear programming and dynamic
A New Optimized Approach to Resolve a Combinatorial Problem … 949
programming, as well as, the genetic algorithms, ant colony optimization and artifi-
cial bee colony optimization wherein the input scale is increasing [9–14]. Further-
more, the aforementioned problem is a complex combinatorial optimization problem
among the NP-Hard problems, that requires exploring new resolution methods,
specifically the bio-inspired metaheuristics. The Cat Swarm Optimization inspired
from real cat behaviors, reveals to be efficient for the TSP resolution [6]. Spider
Monkey Optimization is a new method among SI algorithms, is proved to solve the
TSP with good performances [4]. Indeed, the bio-inspired methods used to solve TSP,
for instance genetic algorithms, ant colony optimization, particle swarm optimization
and bat algorithm optimization, are not necessary providing the best solutions [15].
Differential evolution is an evolutionary algorithm, usually used to solve continuous
optimization problems. A discrete version of the differential evolving algorithm is
proposed to improve resolving of complex discrete TSP [16]. The TSP problem is
described as a salesman who is required to choose an itinerary of minimum distance
to carry out the distribution of its merchandises. The salesman must travel from the
depot, visit each city once and return to the starting point [17]. The enumeration of all
tours for a given problem is not a possibly appropriate method. Corresponding to an
n-city TSP, the possible permutations is equal to (n − 1)! Routes [18]. In addition, the
exact methods are requested to constitute a search space of all possible permutations
of cities, then to find the corresponding optimal solutions. The resulting run time is
a polynomial factor of O(n!), as n is the length of the route. The exact resolutions of
large size problems are inoperable [7]. As the genetic algorithm is well performing
in achieving good solutions for TSP, this genetic approach is also inconvenient in
case of premature convergence, expensive evolving routines and non-performance
of local search techniques. Thus, an immune based genetic algorithm is attested to
exceed these inconvenient and improving TSP solutions [15]. Admittedly, the TSP
is formulated mathematically as following [19]:
n n
Minimize : z = ci j xi j
i=1 j=1
n
Subject to : xi j = 1, j = 1, 2, . . . .n,
i=1
n
xi j = 1, i = 1, 2, . . . .n,
j=1
xi j ∈ {0, 1}, i, j = 1, 2, .......n.
x f or ms a H amiltonien cycle (1)
Where cij and xij are the costs and the decision variables respectively linking
position i to position j. Vector x̃ has the whole sequence of the resulting route which
is a solution for the TSP [20]. In what follows, a literature review is achieved for
studying a Kohonen-type neural network in order to investigate its performances in
resolving TSP.
950 O. El Majdoubi et al.
The geometric tour is modeled as a stretching ring of nodes while the network
is evolving. The input data are presented randomly in the network. The neuron
network competition is based on considering Euclidean distance between nodes of
the network. The winner node J is obtained with the minimum distance to the present
city as formulated below [29]:
J = Argminj xi − yj 2 (2)
Where (xi, yi ) is the coordinate of city i and ||.|| is the Euclidean distance [29]. The
SOM adaptation procedure is applying firstly via an inhibition mechanism in order
to explore a winner node corresponding to each city up to represent all the cities
of the network once in one vector. The winner node is selected when the Euclidean
distance connecting the given city and the candidate node is minimum. Furthermore,
the adaptation procedures result in moving the winner node and its neighbors towards
the presenting city as following [30]:
vj∗ = vj + μf(., .) c − vj (3)
Where G is the gain parameter and d is the cardinal distance measured along the
ring consisted of m nodes. The gain parameter is initialized by G0 = 0.06 + 12.41 m
and updated for each iteration. The predefined values of learning and decreasing rates
are μ = 0.6 and α = 0.1 [30]. The competitive training algorithm is used to allow the
winning neuron to learn as well as the neurons located within a predefined radius from
the winning neuron and to update the related adaptation measures simultaneously
[29]. The cardinal distance measured along the ring between nodes j and J, where m
952 O. El Majdoubi et al.
Importantly, the neighborhood function has an impact on planning for the shortest
path by exploring closer nodes to each other’s [29]. It is important to develop heuris-
tics for parameters adaptation and adjustment in order to reach a faster convergence.
The main adaptation parameters are updated as described on Eqs. 6 and 7. The
learning rate is related to the number of iterations. Furthermore, the neighborhood
function variance is defined depending to an appropriate initialization equal to 10
[29].
1
αk = √
4
(6)
k
The resolution of Euclidean TSP is also made via hybrid methods, for instance,
by combining SOM procedure within an evolutionary dynamic algorithm, by the
integration of a mapping operator, a fitness evaluation and a selection procedure.
This method is interesting justified by the efficiency and the performances relative to
computed solutions and run time [31]. The hybridization of evolutionary algorithms
is a powerful method for the resolution improvement. The recombination of algo-
rithms is possibly made by implementing unrelated methods to previous information
of parents, evolutionary process, or designing interaction models within the set of
solutions [32].
Recently, the world faced a new respiratory virus so-called coronavirus (COVID-
19). At the event of the outbreak, a novel bio-inspired metaheuristic is developed,
modeling the spread of coronavirus. Since an initial infected individual called Patient-
Zero (PZ), the coronavirus infects new individuals according to a specific rate and
create a search space of infected individuals. Each individual belongs to infected
population, is possibly constrained to die, infect or recover. The main properties of
CVOA are as follows: The probabilities and parameters are defined also updated by
scientific community, the exploration of search space is handled as long as the infected
population is not null and the high rate of expansion ensures better use of search space
leading to the intensification of the resolution. Also, the concept of parallel strains
reformulated as applying the processing algorithms in different processors in order to
generate diversification of the resolutions. The Coronavirus optimization algorithm
begins by generating a random initial population since the individual PZ. The next
step concerns the exploration of dead individual depending on die probability(P_DIE)
A New Optimized Approach to Resolve a Combinatorial Problem … 953
and make their removal from the infected population. This step so-called intensifica-
tion is also handled for the detection of recovered individual who possibly can infect
others, depending on probabilities of spreading (P_SUPERSPREADER). Also, for
ensuring diversifications of individuals, it is appropriate to define an immigration
probability (P_TRAVEL) for detecting individuals enabling the spread of coron-
avirus in other solutions. The update of populations is ensured for each iteration.
Each dead individual is appended in dead population to avoid reusing it. After each
iteration, the infected population are sent to recovered population. The recovered
individual can be reinfected if the probability of reinfection (P_REINFECTION) is
acquired. Thus, each individual is subjected to distancing or isolation measures if the
probability of isolation is achieved. Finally, each infected individual is sent to new
infected population. The stop criterion is expected if there is no evolution of the infec-
tion, explained by the decreasing size of infected population and new infected popu-
lation, or if the individuals in the new infected population cannot transmit anymore
the virus. Generally, the stop criterion is achieved throughout a specific duration
called PANDEMIC_DURATION [33]. The main CVOA algorithm [33] is presented
as following (Fig. 3):
and 0.06 related to the Euclidean TSPLIB instance Burma14 and corresponding to
results of Tables 1, 2 and 3 respectively. Also, the average deviation related to the
TSPLIB instance Ulysses16 are 0.002, 0.06 and 0.017 corresponding to SOM resolu-
tion and to hybrid proposed algorithm considering a predefined and adapted learning
parameter respectively. The calculated average deviation related to the TSPLIB
instance Ulysses22 are 0.3, 0.4 and 0.3 resulting since the three considered imple-
mentation tests respectively. The considered approaches presented in this work are
efficient to resolve the TSP corresponding to small and medium data scale. Indeed,
corresponding to the TSPLIB instances namely Berlin52, Att48, Rat99 and Pr76,
the SOM resolution is not good performing. Thus, the results of the aforementioned
instances are better improved by the implementation of the proposed hybrid method
especially, whereas the learning parameters are adapted. The obtained average devi-
ation is 0.3 and 0.5 corresponding to Pr76 and Rat99 respectively, resolved by
the present combined algorithm and where the learning parameters are adapted.
Therefore, it is important to improve the present proposed resolution by tuning of
the considered parameters or experiment the successful neighborhood procedures.
Importantly, it is necessary to apply the techniques of parallel computing in order to
achieve fast convergence and to improve resolutions.
References
9. Durbin R, Willshaw D (1987) An analogue approach to the travelling salesman problem using
an elastic net method. Nature 326:689–691
10. Wang J, Ersoy OK, He M, Wang F (2016) Multi-offspring genetic algorithm and its application
to TSP. Appl Soft Comput 30:484–490
11. Ezugwu AES, Adewumi A, Frîncu M (2017) Simulated annealing based symbiotic organisms
search optimization algorithm for traveling salesman problem. Expert Syst Appl 77:189–210
12. Choong SS, Wong LP, Lim CP (2019) An artificial bee colony algorithm with a modified choice
function for the traveling salesman problem. Swarm Evol Comput 44:22–635
13. Chen WL, Dai SG (2006) Survey of algorithms for traveling salesman problem. J Chuzhou
Univ 8(3):1–6
14. Guo JY (2006) An overview of traveling salesman problem. Pop Sci Technol 8:229–230
15. Lahjouji A, Tajani C, Krkri I, Fakhouri H (2019) Immune based genetic algorithm to solve a
combinatorial optimization problem: Application to traveling salesman problem. In: Ezziyyani
M (eds) AI2SD Conference 2018, Advances in intelligent systems and computing, vol 915.
Springer
16. Ismail MA, Daryl E, Kathryn K (2020) A novel design of differential evolution for solving
discrete TSP. Swarm Evol Comput 52:100607
17. Hoffman KL, Padberg M, Rinaldi G (2013) Traveling salesman problem. In: Gass S I, Fu MC
(eds) Encyclopedia of operations research and management science. Springer, Boston
18. Gilmore PC, Gomory RE (1964) A solvable case of the traveling salesman problem. Proc Natl
Acad Sci 51:178–181
19. Ahuja RK, Magnanti TL, Orlin JB (1993) Network flows. Prentice-Hall
20. Bazaraa MS, Jarvis JJ, Sherali HD (1990) Linear programming and network flows. Wiley, New
York
21. Koprinkova HP, Mladenov V, Kasabov NK (2015) Artificial neural networks. Springer Series
in Bio-/Neuroinformatics
22. Trick M (2008) Oper Res Lett 36(2):276–277
23. Khourdifi Y, Bahaj M (2020) Analyzing social media opinions using hybrid machine learning
model based on artificial neural network optimized by particle swarm optimization. In:
Ezziyyani M (eds) AI2SD Conference 2019. Advances in Intelligent Systems and Computing,
vol 714. Springer, Singapore
24. Kohonen T (1990) The self-organizing map. Proc IEEE 78(9):1464–1480
25. Kohonen T (2001) Self-organizing maps. Springer , Heidelberg
26. Somhom S, Modares A, Enkawa T (1997) A self-organising model for TSP problem. J Oper
Res Soc 48(9):919–928
27. Kohonen T, Kaski S, Lagus K, Salojarvi J, Honkela J, Paatero V, Saarela A (2000) Self
organization of a massive document collection. IEEE Trans Neural Netw 11(3):574–585
28. Moukhafi M, Yassini KE, Bri S, Oufaska K (2019) Artificial neural network optimized by
genetic algorithm for intrusion detection system. In: Ezziyyani M (eds) AI2SD Conference
2018. Advances in Intelligent Systems and Computing, vol 714. Springer, Singapore
29. Yanping B, Wendong Z, Zhen J (2006) A new self-organizing maps strategy for solving TSP.
Chaos Solitons Fractals 28(4):1082–1089
30. Jan F, Miroslav K, Vojtěch V, Libor P (2011) An application of the self-organizing map in the
non-euclidean TSP. Neurocomputing 74(5):671–679
31. Créput JC, Koukam A (2009) A memetic neural network for the euclidean TSP. Neurocom-
puting 72(4–6):1250–1264
32. Tinós R (2020) Artificial neural network based crossover for evolutionary algorithms. Appl
Soft Comput 95:106512
33. Martínez-Álvarez F, Asencio-Cortés G, Torres JF, Gutiérrez-Avilés D, Melgar-García L, Pérez-
Chacón R, Rubio-Escudero C, Riquelme JC, Troncoso A (2020) Coronavirus optimization
algorithm: a bioinspired metaheuristic based on the COVID-19 propagation model. Big data
8(4):308–322
Study of a Dual-Function Intelligent
Electronic Pin to Help Compliance
with Security Measures Related
to Covid-19
Laince Pierre Moulebe, Abdelwahed Touati, Eric Obar Akpoviroro,
and Nabila Rabbah
Abstract COVID-19 is a virus that developed in December 2019, the disease has
spread around the world and there are many cases per day. The latest information
reports that there are more than sixty-nine million cases of infection worldwide
according to an official diagnosis. The new coronavirus pandemic has killed at least
one million six peoples worldwide since its emergence, according to a report by AFP
(Agence France Presse), but the actual number estimate higher, knowing that some
countries are exceeded by the number of cases in hospitals and that others do not
have the resources to test the majority of their population. Several treatment methods
to slow down and eliminate Covid-19 exist to date. Indeed, following the appearance
of the new Corona Virus, the world scientific community has carried out several
research and development for eliminated the virus, there are smart masks, respirators,
gels, mobile applications, vaccines that are gradually developing and much more
research. However, according to studies, transmission is mainly by direct contact,
including contact of uninfected hands with the face, non-compliance with distancing
measures. And, according to psychologists, one of the most effective methods to
fight against the Covid-19 is to prevent these frequent forget fulnesses. On the other
hand, research shows that hand-to-face contact is a frequent human behavior. Hence,
the main objective of this article is to propose a prevention technique through the
study of an electronic device that allows users to keep their safety distance, while
reminding them of the approach of a hand towards their face. It is an electronic pin
featuring PIR sensors, with lenses designed to detect at distances 1.5 m. Containing
a battery for power, a button, which turn on or off the pin, depending on whether you
are in a bus for example or in an airport. Besides, this simple electronic component
designed, pin will have suitable cost for all social categories.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 959
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_87
960 L. P. Moulebe et al.
1 Introduction
Covid-19 was declared by the World Health Organization on March 11, 2020 as
a pandemic following the number of deaths which has multiplied 13 times more
than that recorded in China [1]. Covid-19 continues to increase today, with more
than sixty-nine million cases positive infection that have been officially diagnosed in
208 countries and territories since the start of the epidemic. According to the AFP,
the actual number of contaminations is much higher than this value, knowing that a
large number of countries now only testing cases requiring hospital treatment. Among
these cases, at least forty-four million are now considered cured. Covid-19 is usually
spread through close contact (shaking hands, by the airway, coughing and sneezing).
Several researches are established in order to fight this pandemic [2], through some
suggestions for health workers because they are less exposed and their availability
is essential [3, 4], moreover according to [5] it is important to take an interest in
the behavior of people working in hospitals, on their knowledge about the epidemic.
Despite the containment including about 4.5 billion of the population, awareness of
behaviors and habits, respect for hygiene, we note that the number of cases continues
to increase. Practical innovative systems such as automatic disinfectants suitable for
entering large spaces [6], smart bracelets allowing hands to be disinfected at any time
[7], it is clear that innovations are continually emerging. All these systems aim to
fight against the pandemic which affects more than 300 people every minute [8]. To
date several countries have developed vaccines which are being distributed [9–11],
however populations remain hesitant about these vaccines [12, 13], because some
of the first vaccines were not effective from the start. This distrust of the population
shows that it remains important to offer solutions in addition to vaccines which are
the subject of speculation, this is why in this article we propose another innovation in
the prevention of Covid-19. According to [14], touching the face is a typically human
behavior that expresses self-awareness; however, it is one of the best ways to prevent
coronavirus spread. Several studies on self-contact have been established over the
time [15–19], showing that, it is a behavior that is repeated frequently, but further
research shows that in self-contact the contact of hand to face is the most frequent
[20]. In [21] a study shows the behavior concerting the contact of the hands with
the face to perform on 60 peoples. We see that it is indeed one of the most frequent
human behaviors, but the best way to fight covid-19 is to limit it, even avoid contact
of non-disinfected hands with the face because each forget (Fig. 1) may be likely to
cause multiple casualties. Hence, to limit the errors and forgetting when faced with
Covid-19 concerning touching the face, in this article we propose the study of an
electronic device that would allow social distancing (Fig. 2) while allowing users
to always clean their hand before touching their face and keep the distancing. The
electronic system studied mainly consists of a PIR sensor, a lens, an on/off button,
a power supply battery, a microcontroller and a sound. The studied device will be
worn like a pin, its size is around 50 mm.
Study of a Dual-Function Intelligent Electronic Pin … 961
This work is divided into three parts, the first paragraph presenting the description
of the system, then in the second the proposed technical solution followed by a
conclusion.
2 Description Studied
Devices used to remind humans to perform tasks are found a lot in our environment,
in the health sector we use connected objects such as Medi’Rappel, Medisafe, to
remind to take these drugs or to respect the dosage, warning lights used in cars,
control systems in industry. The Covid-19 is mainly transmitted between people by
respiratory droplets and contact [21–25],In an analysis of 75,465 cases of Covid-19
in China, no transmission by air was reported [26]. The device is designed to detect
peoples over a distance of 0 to 1.5 m, the device mainly consists of a PIR (Fig. 3)
detector which is ideal motion sensors for humans and animals [27], a battery, a
microcontroller allowing the sounder to be put on and standby for 5 after detection
of the approach of a hand towards the face or of a person within a radius of 1.5 m
according to the simple algorithm (Fig. 4) and a microphone (Fig. 5 b). Note that
these are miniaturized devices with the smallest sizes. The detection distance is an
important parameter who is calibrate with the lens of the PIR sensor. Through the
work established by [27, 28], we can evaluate the characteristics (Fig. 5 a) of the
battery for a detection distance between 0–1,5 m which is around 5 V. The electronic
pin is designed to allow users to keep their safe distance. Therefore, the user wearing
the pin is automatically alerted by a song in the event of presence of a person within
1.5 m, but also when a hand approaches the face, knowing that the hand passes
through the field of the lens, this sensor can be shut down if we are in a bus or public
transport where compliance with the measures is difficult.
In this flowchart we show the system that puts the sensor on stand-by, (1) the
sensor is active and detects a hand, then the alert is issued and (2) the user cleans
his hands. Subsequently, through a microcontroller, the sensor is put on stand-by for
5 min (3). And finally the controller activates the sensor after 5 min (4).
962 L. P. Moulebe et al.
3 Technical Solution
Alert
4 sensor on
4 Conclusion
In this work we were able to highlight the fact that COVID-19 is mainly spread by
contact, and that despite several restriction such as containment, awareness to clean
hands and safety distance, the number of cases continues to increase, then it has been
shown that touching the face is a typically human behavior. This allowed us to show
the need for a simple system allowing users to have a simple and easily portable
system that reminds them to clean their hands before touching their face but also to
respect social distancing measures.
Thus, omissions of safety measures in supermarkets, in front of banks or any other
space will be limited as well for contact of the hands with the face. In definitive this
solution will help to reduce the propagation of the Covid-19.
Study of a Dual-Function Intelligent Electronic Pin … 965
References
1. Ho LTF, Chan KKH, Chung VCH, Leung TH (2020) Eur J Integr Med 101116. https://doi.org/
10.1016/j.eujim.2020.101116
2. Yen M, Schwartz J, Chen S, King C, Yang G, Hsueh P (2020) ScienceDirect Interrupting
COVID-19 transmission by implementing enhanced traffic control bundling: implications for
global prevention and control efforts. J Microbiol Immunol Infect. 3. https://doi.org/10.1016/
j.jmii.2020.03.011
3. Xu K, Lai X, Zheng L (2020) Suggestions on the prevention of COVID-19 for health care
workers in department of otorhinolaryngology head and neck surgery. World J Otorhinolaryngol
Neck Surg. https://doi.org/10.1016/j.wjorl.2020.03.002
4. Tebala GD, Bond-Smith G ( 2020) Guidelines and recommendations during the COVID-19
pandemic: a word of caution. Am J Surg 220(6):1526–1527. https://doi.org/10.1016/j.amjsurg.
2020.06.005
5. Shi Y et al (2020) Brain, Behavior, & Immunity - Health Knowledge and attitudes of medical
staff in Chinese psychiatric hospitals regarding COVID-19. Brain Behav Immun Heal 4:100064.
https://doi.org/10.1016/j.bbih.2020.100064
6. Hussain S et al (2020) Proposed design of walk-through gate (WTG): Mitigating the effect of
COVID-19. Appl Syst Innov 3(3): 1–7. https://doi.org/10.3390/asi3030041
7. El Majid B, Motahhir S, El Hammoumi A, Lebbadi A, El Ghzizal A Preliminary design of a
smart wristband disinfectant to help in Covid-19 Fight. https://doi.org/10.3390/inventions50
30032
8. CovInfo - Suivez l’évolution du Coronavirus dans le monde avec notre bulletin quotidien.
https://www.covinfo.fr/monde. Accessed 06 Dec 2020
9. Reddy KA, Srinivas K, Ayyappan GS (2015) Simulation of power management circuit in a
standalone solar photovoltaic-fuel cell-battery hybrid power system. In 2015 IEEE IAS Joint
Industrial and Commercial Power/Petchem Product Industries ICPSPCIC, 168–173. https://
doi.org/10.1109/CICPS.2015.7974070
10. “Covid-19 vaccine: who are countries prioritising for first doses? | World news | The
Guardian.” https://www.theguardian.com/world/2020/nov/18/covid-19-vaccine-who-are-cou
ntries-prioritising-for-first-doses. Accessed 11 Dec 2020
11. Vaccins contre le Covid-19 : quels sont les pays qui ont déjà passé commande ?
- L’Express.” https://www.lexpress.fr/actualite/monde/vaccins-contre-le-covid-19-quels-sont-
les-pays-qui-ont-deja-passe-commande_2138952.html. Accessed 11 Dec 2020
12. On Kwok K et al (2020) Journal Pre-proof Influenza vaccine uptake, COVID-19 vaccination
intention and vaccine hesitancy among nurses: a survey TITLE Influenza vaccine uptake,
COVID-19 vaccination intention and vaccine hesitancy among nurses: a survey. Int J Nurs
Stud (2020). https://doi.org/10.1016/j.ijnurstu.2020.103854
13. Vaccine hesitancy in the University of Malta Faculties of Health Sciences, Dentistry and
Medicine vis-à-vis influenza and novel COVID-19 vaccination | Elsevier Enhanced Reader.
https://reader.elsevier.com/reader/sd/pii/S0378378220307623?token=22F51A1FB4FEFC9
15C562BAC3928A46C9FB5C26CBB714C789EB335A44A2423C702966C5AD8E77760B
4D65FC3F6264393. Accessed 07 Dec 2020
14. Why you can’t stop touching your face, even though it’s one of the best ways to prevent coron-
avirus spread. https://www.businessinsider.fr/us/psychological-reason-why-you-cant-stop-tou
ching-face-coronavirus-2020-3?utm_source=feedburner&utm_medium=referral. Accessed
09 Apr 2020
15. Ekman P, Friesen WV (1974) Nonverbal behavior and psychopathology. In: Kats MM,
Friedman RJ (eds) The psychology of depression. Contemporary theory and research, New
York, pp 203–224
16. Barroso F, Freedman N, Grand S (1980) Self-touching, performance, and attentional processes.
Percept Mot Skills 50(3):1083–1089
17. Hampson E, Kimura D (1984) Hand movement asymmetries during verbal and nonverbal tasks.
Canadian J Psychol 38(1):102–125 PMID: 6713294
966 L. P. Moulebe et al.
18. Harrigan JA (1982) (1985) Self-touching as an indicator of underlying affect and language
processes. Soc Sci Med 20(11):1161–1168
19. Leavens DA, Aureli F, Hopkins WD, Hyatt CW (2001) Effects of cognitive challenge on self-
directed behaviors by chimpanzees (Pan troglodytes). Am J Primatol 55(1):1–14. https://doi.
org/10.1002/ajp.1034. PMID: 11536312
20. Goldberg S, Rosenthal R (1986) Self-touching behavior in the job interview: antecedents and
conse- quences. J Nonverbal Behav 10(1):65–80
21. Mueller SM, Martin S, Grunwald M (2019) Self-touch: contact durations and point of touch of
spontaneous facial self-touches differ depending on cognitive and emotional load. PLoS One
14:3. https://doi.org/10.1371/journal.pone.0213677
22. Liu J, et al. (2020) Community Transmission of Severe Acute Respiratory Syndrome Coro-
navirus 2, Shenzhen, China, 2020. Emerg Infect Dis 26(6). https://doi.org/10.3201/eid2606.
200239
23. Modes of transmission of virus causing COVID-19: implications for IPC precaution recom-
mendations https://www.who.int/news-room/commentaries/detail/modes-of-transmission-of-
virus-causing-covid-19-implications-for-ipc-precaution-recommendations. Accessed 11 Apr
2020
24. Li Q et al (2020) Early transmission dynamics in Wuhan, China, of novel coronavirus-infected
pneumonia. N. Engl J Med 382(13):1199–1207. https://doi.org/10.1056/NEJMoa2001316
25. Chan JFW et al ( 2020) A familial cluster of pneumonia associated with the 2019 novel
coronavirus indicating person-to-person transmission: a study of a family cluster. Lancet
395(10223):514–523. https://doi.org/10.1016/S0140-6736(20)30154-9
26. Frieden TR, Lee CT (2020) Identifying and Interrupting Superspreading Events—Implications
for Control of Severe Acute Respiratory Syndrome Coronavirus 2. Emerg Infect Dis. 26(6)
https://doi.org/10.3201/eid2606.200495
27. Gami H (2017) Movement direction and distance classification using a single PIR sensor. IEEE
Sensors Lett. 2(1):1–4. https://doi.org/10.1109/lsens.2017.2782179
28. Zappi P, Farella E, Benini L (2010) Tracking motion direction and distance with pyroelectric
IR sensors. IEEE Sens J 10(9):1486–1494. https://doi.org/10.1109/JSEN.2009.2039792
DenTcov: Deep Transfer Learning-Based
Automatic Detection of Coronavirus
Disease (COVID-19) Using Chest X-ray
Images
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 967
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_88
968 Y. El Idrissi El-Bouzaidi and O. Abdoun
2 Related Work
Since computer-assisted image and deep learning technologies have proven their
superiority for medical image analysis, with the reliability to detect COVID-19 into
chest X-ray image [6, 7], researchers use advances of these techniques as a tool to
diagnose COVID-19. Applying artificial intelligence offered significant advantages
in detecting the disease and measuring infection rates, with promising results.
DenTcov: Deep Transfer Learning-Based Automatic Detection .. 969
Farooq and Hafeez [8] proposed by successive stages based on the architecture of
ResNet-50 a model, called COVIDResNet, using a collection of radiographic image
datasets that classifies patients as normal, bacterial, viral Pneumonia, and COVID-19
positive infections. An accuracy of 96.23% was achieved.
In another Paper [9], transfer learning techniques applied with the concatena-
tion of pre-trained models, ResNet50V2 and Xception, the proposed approach was
compared to Xception Res-Net50V2 models as a means of classifying chest X-ray
images according to the normal, COVID-19, and non-COVID pneumonia classes.
They used five-fold cross-validation as part of their classification testing process.
This technique resulted in approximately an accuracy of 91.4%.
Narin et al. [10] showcased several deep learning approaches, including three
distinct deep learning architectures, ResNet-50, Inception-ResNetV2, and Incep-
tionV3, The binary classification is used to discriminate COVID-19 distinct to the
healthy image. The evaluation indicates that ResNet50 had the highest classification
efficiency with an accuracy of 98.0%, recall of 96%, and accuracy of 100%, against
the accuracy of 97.0 and 87% for InceptionV3 and Inception-ResNetV2, respectively.
Chowdhury et al. [7] demonstrate some techniques with a transfer technique
used to detect an infection caused by COVID-19 on X-ray images of the lungs. They
demonstrate using pre-trained architecture such as AlexNet, DenseNet-201, ResNet-
18, and SqueezeNet, employed as classifiers. A classification accuracy of 98.3% was
achieved using the SqueezeNet network, with a specificity of 99%.
Asif and Wenhui [11] propose a system to automatically diagnose COVID-19,
the Transfer Learning Based Inception V3 was used to diagnose the patient’s chest
infection. This model was tested against normal radiographic imaging, pneumonia,
and COVID-19. Classification accuracy of 96% was achieved.
Khan et al. [12] have come with a classifier of chest X-ray images called CoroNet,
classifying images as Normal, Bacterial, Viral Pneumonia, and COVID-19 based on
the Xception architecture. As a result, they obtained an overall performance accuracy
of up to 89.6%.
Sethy and Behera [13], a number of various CNN models for extraction features
were trained using radiographic imagery of different architectures together in
conjunction with the Support Vector Machine for detecting COVID-19 positive cases,
thus achieving the accuracy of 95,38%.
In this section, the DenTcov model is illustrated and described in detail. The archi-
tecture of the different parts is illustrated (see Fig. 1). The model was defined in two
phases. Phase I consists of (1) pre-processing and (2) data augmentation. In phase
II, (1) a deep convolutional network is specified as the first step, (2) the best deep
transfer learning model has been defined in the second step. a background discussion
on the process for each part of the model is given.
970 Y. El Idrissi El-Bouzaidi and O. Abdoun
Train Set
Validation Set
Test Set Automatic Learning rate
Train Loss
Update
Pre-Processing Augmentation
Input
X-ray
images
(COVID-19 Probabilities)
Fig. 2 A Representative chest x-ray images after performing the Augmentation techniques
Data augmentation has occurred in the creation of new data using modifications made
to the dataset. Basically, with reasonable modifications and by applying different
transformations randomly to our training set, we create new augmented data (see
Fig. 2).
The following Table 1 summarizes the transformations we used in this work.
For the purpose of this paper, for feature extraction, the knowledge from an existing
set of Convolutional Neural Network architectures pre-trained was used, with excel-
lent results achieved by using a variety of classification tasks as alternatives to
proposing our own architecture from scratch. For this reason, The DensNet121 archi-
tecture [17] has been used as an effective way for feature extraction by using its
weights that are pre-trained in the ImageNet dataset [15] using reduced parameters.
Another reason to use this network is to take advantage of the network’s condensed
structure, offering easy training, especially due to the reuse of features by other
layers, which increases the variety of its content between layers and improves the
network’s performance and generalization. we also adopt advanced high-speed tech-
niques introduce in Fastai [16], allowing us to improve the accuracy of our model
through every few epochs. The subsequent section provides a short summary of the
DenseNet121 architecture.
DensNet121. Transmits the characteristics to all the next layers in each dense block.
In its architecture, DensNet121 is defined as follows: there are four dense blocks,
972 Y. El Idrissi El-Bouzaidi and O. Abdoun
each dense block has a different number of layers (6, 12, 24, 16 consecutive layers).
Between each dense layer, there are transition layers, which allow reducing the size of
the features that are generated by the concatenation. The classification layers contain
a Global Average-Pooling plus a Fully Connected layer. For details please see [17].
Pre-trained Model. Our DenTcov has two components: the first consists essen-
tially of exploiting the capabilities offered by the DensNet121 for the transformation
of visual elements into feature vectors, while the second component consists of a
fully connected layer whose primary objective is to predict classification. As part
of this work, we have added a sequence of layers to the early portion of a typical
DensNet121, namely an adaptive Average/Max pooling, Batch normalization, and
drop out according to [16].
The second part of the DensNet121 is replaced with a fully connected network
so that the probabilities are calculated according to our classes, Normal, Pneumonia,
and COVID-19. As a predictable result, between each of these classes, is the one
having the highest probability. During the training phase, the modified parts of the
model were fine-tuned with pre-trained weights, leaving the original part untrained.
We obtain the best convolutional neural network architecture for transfer learning
and its hyperparameter by training the model and selecting an optimal learning rate
using the cyclic learning rate technique proposed in [18]. We configure the training
process of our model with the parameters given in the following Table 2.
Table 2 Hyperparameters
Parameters Value
details used in DenTcov
model. Batch size 64
Optimizer Adam, SGD
Momentum Betas = (0.9, 0.99)
Classes 3 Classes (Normal, Pneumonia,
Covid-19)
Loss function Crossentropy. Useful in
multi-classification problems
Training epochs 15
Dropout 0.3
Classifier Softmax (layer outputs the
probability distribution)
Activation function ReLu
DenTcov: Deep Transfer Learning-Based Automatic Detection .. 973
4.1 Experiments
Dataset. The dataset was obtained from GitHub, the public source repository that
Dr. Joseph et al. has shared [19] which contains chest X-rays of patients infected by
COVID-19, also X-rays belonging to other COVID 19 families, such as ARDS,
SARS, and MERS. In our dataset, we only included radiographic images with
COVID-19, representing a total of 190 images, 264 normal x-ray images, and 275
bacterial and viral pneumonia infections, obtained from the Kaggle “Chest XRay
Images (Pneumonia)” [20].
Training. During the training, only the newly added layers of the network and last
layers group; Fully connected layers, are trained for 5 epochs while preserving the
weights of ImageNet for the rest of the network. We use the cyclic learning rate
technique introduced in [18], to help select the optimal learning rate. Using these
discriminatory learning rates, which keep a low learning rate for the initial layers
because they need less tuning and progressively increase the learning rate on the later
layers that need more tuning, especially those that are fully connected. By using the
learning rate (Max_lr) in the range 1e−4, 1e−3, all the network is fine-tuned, as first
layers with 1e−4 and the last layers with 1e−3. Each intermediate layer has been
trained with learning rates between these values for 10 epochs.
Evaluation Metrics. We have provided the four evaluation matrices to judge the
achievement of the results obtained from the DenTcov model. The true positives,
true negatives, false positives, and false negatives correspond to T.P, T.N, F.P, and
F.N, respectively, which are as follows in Table 3.
4.2 Results
The DenTcov was evaluated for two scenarios. The first is the classification of the
radiographic images into three classes: pneumonia, COVID-19, and Normal. The
second scenario is the classification of both COVID-19 and Pneumonia.
Figures 3, 4 illustrates the curves representing the loss convergence relating to
the training and the validation phases for 3-class (see Fig. 3) and 2-class (see Fig. 4),
respectively, the loss function going into 70 epochs and the highest classification
achieved an Accuracy of 96.52% (see Fig. 5).
Figure 6 gives an overview of the confusion matrix of the COVID-19 class
compared to the other classes.
We show how the Pneumonia class is not classified well (see Fig. 6), This problem
may be due to the fact that the class group two types of pneumonia: Viral and
Bacterial. Moreover, people who develop Pneumonia can have different levels of
progression. It is also notable that the classes COVID-19 achieves a perfect sensitivity,
and Class Normal already have a good level of correct classification. An excellent
compromise between the true and predicted values by DenTcov; which made it
possible to verify the effectiveness of the model, these results provide an encouraging
Loss
Batches processed
Batches processed
Error_rate
Accuracy
Covid-19 Covid-19
Actual
Actual
Normal
Pneumonia
Pneumonia
Covid-19
Pneumonia
Covid-19
Pneumonia
Normal
Predicted Predicted
Fig. 6 The confusion matrix of DenTcov for 3-class & 2-class classification
as a tool of reducing stress on the health care system. Table 4 details the precision,
the recall, and the F1 score, respectively, for each class.
4.3 Discussion
Our current work focuses on automated detection from X-ray images of the charac-
teristics that could be significantly related to COVID-19 disease compared to other
respiratory diseases, in order, to classify it effectively, the use of the DensNet121
architecture was effective in this task, due to the high accuracy that has been achieved,
96.52 and 99% for three and two classes, respectively. Another key advantage, getting
a higher recall value signifies fewer False Negative, which is a significant result in
976 Y. El Idrissi El-Bouzaidi and O. Abdoun
5 Conclusion
References
7. Chowdhury MEH et al (2020) Can AI help in screening Viral and COVID-19 pneumonia?
IEEE Access 8:132665–132676
8. Farooq M, Hafeez A (2020) COVID-ResNet: A Deep Learning Framework for Screening of
COVID19 from Radiographs. arXiv:2003.14395 [cs, eess]
9. Rahimzadeh M, Attar A (2020) A modified deep convolutional neural network for detecting
COVID-19 and pneumonia from chest X-ray images based on the concatenation of Xception
and ResNet50V2. Inform Med Unlocked 19:100360
10. Narin A, Kaya C, Pamuk Z (2020) Automatic Detection of Coronavirus Disease (COVID-19)
Using X-ray Images and Deep Convolutional Neural Networks. arXiv:2003.10849 [cs, eess]
11. Asif S, Wenhui Y, Jin H, Tao Y, Jinhai S (2020) Classification of COVID-19 from Chest X-ray
images using Deep Convolutional Neural Networks. medRxiv 2020.05.01.20088211. https://
doi.org/10.1101/2020.05.01.20088211
12. Khan AI, Shah JL, Bhat M (2020) CoroNet: A deep neural network for detection and diagnosis
of COVID-19 from chest x-ray images. Comput Methods Programs Biomed 196:105581
13. Sethy PK, Behera SK, Ratha PK, Biswas P (2020) Detection of Coronavirus Disease (COVID-
19) Based on Deep Features and Support Vector Machine
14. Du SS, Koushik J, Singh A, Poczos B (2017) Hypothesis transfer learning via transformation
functions. In Guyon I, et al (eds) Advances in Neural Information Processing Systems, vol 30,
pp 574–584. Curran Associates, Inc.
15. Deng J, et al (2009) ImageNet: A large-scale hierarchical image database. In: 2009 IEEE
Conference on Computer Vision and Pattern Recognition 248–255 (2009). https://doi.org/10.
1109/CVPR.2009.5206848
16. Howard J, Gugger S (2020) Fastai: a layered API for deep learning. Information 11:108
17. Huang G, Liu Z, Van Der Maaten L, Weinberger KQ (2017) Densely Connected Convolutional
Networks. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR),
pp 2261–2269. https://doi.org/10.1109/CVPR.2017.243
18. Smith LN (2017) Cyclical Learning Rates for Training Neural Networks. arXiv:1506.01186
[cs]
19. Cohen JP (2020) ieee8023/covid-chestxray-dataset (2020). Accessed 02 Oct 2020
20. Chest X-Ray Images (Pneumonia) | Kaggle. https://www.kaggle.com/paultimothymooney/
chest-xray-pneumonia. Accessed 02 Oct 2020
21. Ozturk T et al (2020) Automated detection of COVID-19 cases using deep neural networks
with X-ray images. Comput Biol Med 121:103792
Network Technology
Explainable Deep Learning Model
for COVID-19 Screening in Chest
CT Images
1 Introduction
COVID-19 is an infectious and contagious virus and, given its global magnitude, the
World Health Organization (WHO) has announced it as a pandemic. The declaration
of a pandemic also highlighted the deep concern about the dangerous speed of spread
and severity of COVID-19. To date (mid-November), the number of patients has
exceeded 54 million worldwide.
Worldwide, the most effective screening technique is Reverse Transcription Poly-
merase Chain Reaction (RT-PCR). Nevertheless, the poor sensitivity of RT-PCR
and the lack of test kits in pandemic regions increases the workload of screening,
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 981
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_89
982 M. El Habib Daho et al.
so numerous sick persons are not immediately confined [14] and this accelerates
the spread of COVID-19 as many infected subjects cannot obtain immediate medi-
cation. In this circumstance, it is important to use other diagnostic techniques to
better control the pandemic. Healthcare departments recommend the use of medical
imaging (computed tomography (CT) and chest X-Ray) as a complement to RT-
PCR. Recent studies have shown that CT has a high diagnostic and prognostic value
for COVID-19. Computed Tomography is an easy exam and quick to obtain at no
additional cost and has shown a much higher sensitivity than RT-PCR for the inves-
tigation of COVID-19 [9]. The CT-scan image can distinguish potential high-risk
subjects that are predisposed to become critical and require instant medical support.
With advances in artificial intelligence, machine learning (ML), and computer
vision, the build of smart prediction and diagnostic models has grown at an acceler-
ated velocity [11]. Improving these intelligent models for disease prediction and diag-
nosis requires the use of ML techniques [9, 14]. However, to get better ML models,
efficient and effective feature extraction techniques are needed. Deep learning (DL),
as an ML method, has given encouraging performances in facilitating the diagnosis
of lung diseases using CT-scan images of the chest [2, 13]. Deep learning, unlike the
classic ML, can extract relevant characteristics automatically without expert feature
engineering. In this work, we introduce a diagnostic support system for COVID-19
by classifying CT-scan images.
Since the appearance of the COVID-19 pandemic, several works on the classifi-
cation and analysis of scanner images have been carried out. Among these works,
researchers such as [16] and [1] have opted to use classic machine learning tech-
niques. In Barstugan et al. [1], the authors proposed several feature extraction tech-
niques based on the Grey Level (GL Co-occurrence Matrix, GL Run Length Matrix,
GL Size Zone Matrix), Discrete Wavelet Transform, and Local Directional Pattern.
The classification task was performed using SVM.
Other researchers have used deep learning techniques for this task. Among these
works, we find the work of Jin et al. [7], Xu et al. [22] and Wang et al. [21]. They
have used CNN models (Convolutional Neural Networks) as DL techniques for the
screening of COVID-19 using CT-scan images. Pathak et al. [12], Ying et al. [19]
and Li et al. [10] have performed a transfer learning on the ResNet-50 architecture.
Zheng et al. [24] and Jin et al. [8] have used U-Net and CNN as deep learning models
to achieve good performances.
Works cited above focus on classifying COVID-19 images using either machine
or deep learning techniques trained on public datasets. This work aims to propose an
explainable deep learning model for the classification of COVID-19 CT-scan images.
For that, we have collected a new CT-scan dataset for the CHU of Tlemcen in Algeria.
We have also proposed an explainable deep learning classifier based on the Resnet50
architecture and the GradCam algorithm.
Explainable Deep Learning Model for COVID-19 Screening in Chest CT Images 983
2 Methods
Deep learning methods have proven their effectiveness in several fields such as
Computer Vision (CV) and Natural Language Processing (NLP). Deep Convolu-
tional Neural Networks, or CNN for short, are the most popular networks for various
computer vision tasks. These models are reputed to be very good for visual recog-
nition since once a segment in a particular area of an image is learned, the CNN can
recognize that segment anywhere else in the image. Among the commonly used CNN
architectures for the classification of images, we can find VGG16, Resnet50, Incep-
tionV3, and Densenet. We have chosen to use these architectures in this work since
they have been very successful in the challenges of ImageNet and it was proven that
using these pre-trained models in other classification problems (transfer learning)
gives good results.
2.1 VGG
Simonyan and Zisserman [17] proposed the VGG architecture. This model is simple;
it uses only 3 × 3 stacked convolutional layers. The reduction of the convolution
matrix dimension is achieved by a pooling operation (Max Pooling). The fully
connected layers are the two last layers, containing 4,096 neurons each, followed
by a softmax function. VGG16 and VGG19 are the two variants of this architecture.
The “16” and “19” represent the number of layers in the model.
2.2 ResNet
As image recognition becomes more complex, Deep Learning training becomes very
difficult, since extra deep layers are needed to calculate and enhance the accuracy of
the model. Residual training is a concept created to solve this issue and the resulting
model is known as ResNet [5]. ResNet consists of several residual modules where
each module represents a layer. The depth of a ResNet can vary considerably, the
most used are Resnet50 [5] and the one developed by Microsoft which contained 152
layers [5]. Such depth was infeasible before the appearance of ResNet because of the
problem of vanishing gradients. Therefore, ResNet solves this problem by providing
skip connections, applied to each layer before the ReLu activation function, allowing
to preserve the gradient [4].
984 M. El Habib Daho et al.
2.3 Inception
The first version of this model was named GoogLeNet [20], but later variants were
simply denominated Inception. Google published successively new versions of this
architecture. This model uses some convolution filters (1 × 1, 3 × 3, and 5 × 5) in
the same inception module to extract multi-level features.
2.4 DenseNet
DenseNet [6] is composed of blocks called “dense”. In these blocks, each layer is
connected with all the previous layers. To summarize, the DenseNet architecture
makes maximum use of the residual mechanism by making each layer (in the same
dense block) connect to its following layers.
The proposed model used the ResNet50 architecture as shown in the Fig. 1. We have
performed a transfer learning of the ResNet50 model using the ImageNet weights.
Then we added a GlobalAveragePooling layer and two Dense layers of size 1024
and 512. The last classification layer contains two neurons and uses the sigmoid
activation function. For better interpretability of the results, we used the GradCam
algorithm to explain the decision of the model.
3.1 Dataset
In this study, we have used three different datasets; two of them are available for free
access on the internet and the third one we collected from the Hospital of Tlemcen
(CHU Tlemcen).
Yang et al. [23] published the first dataset; it contains 349 COVID and 397 non-
COVID images. This database has been collected from several articles published on
medRxiv, bioRxiv, NEJM, JAMA, Lancet, etc.
The second one is also public and has been published by Soares et al. [18], it
contains 1252 COVID and 1230 non-COVID images. This dataset is gathered from
cases inside hospitals in Sao Paulo, Brazil.
We have collected the third dataset from the Hospital of Tlemcen in Algeria; it
contains 180 COVID and 180 non-COVID images. For training and testing, we used
the three datasets with data augmentation. During the test, we validated the different
models on 355 COVID-19 and 363 Non-COVID images.
Figure 2 shows examples of images with and without the presence of COVID-
19 in the three datasets. On the figure, we can see that the three datasets present
variations in color and texture, which will allow the model to be more robust to the
difference in quality and type of the image presented during the test phase.
3.2 Experiments
With:
True positive (TP): Prediction is COVID and the image is COVID.
True negative (TN): Prediction is Non COVID and the patient is healthy.
False positive (FP): Prediction is COVID and the patient is healthy.
False negative (FN): Prediction is Non COVID and the image is COVID.
Table 1 shows the obtained performances of different models. We can notice that
our approach gives better results compared to the other architectures on accuracy,
specificity, and F-score. For sensitivity, our model is ranked second after Inception
v3, which is good since it is first on the other metrics. In medical diagnostic aid in
general and decision support for COVID-19 in particular this metric is very impor-
tant. Indeed, high sensitivity is equivalent to a very low false-negative rate. False
negatives are patients who have been diagnosed as Non-COVID by the system but
are positive. Such an error can cause the death of the patient and does not have the
Table 1 Results
Accuracy % Sensitivity % Specificity % F-score %
VGG 16 96.10 96.90 95.31 96.08
VGG 19 95.96 94.92 96.96 95.87
DenseNet 121 97.35 97.18 97.52 97.32
Inception V3 97.49 97.74 97.24 97.47
ResNet 50 96.23 96.33 96.14 96.20
Our work 97.63 97.46 97.79 97.60
988 M. El Habib Daho et al.
same impact as a false positive error that can be quickly corrected with additional
tests. We can therefore deduce that our model and Inception V3 are the best in terms
of performance.
Explainable Deep Learning Model for COVID-19 Screening in Chest CT Images 989
4 Conclusion
References
1 Introduction
Vehicular Ad hoc Networks are a subclass of MANETS, and they are considered as
one of the components of Intelligent Transport Systems (ITS). These systems aim to
incorporate new information and communication technologies in the (transportation
field) to improve the safety and the comfort of the road users. A VANET is an ad hoc
type network. It can opportunistically use communications with infrastructures, thus
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 991
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_90
992 S. Achchab et al.
allowing access to other networks and also to the Internet. The various infrastructures
used within the framework of the VANETs are called “Roadside Units” (RSU). These
units can be traffic lights, parking meters, or any other terminal placed at the edge of
the road.
Network simulation [1, 2] in telecommunication and computer networks is a
technique in which the behavior of a real network is modeled using a simulator.
Network simulation tools [1, 2] are used to model real networks with great ease.
There are several network simulators (open source or commercial), which have
different functions [3], such as NS-3 [4], NS-2 [5], OPNET, QualNet, OMNeT,
J-Sim [6] and NetSim.
The NS-3 is one of the most used and popular simulators. The main objective of
this NS-3 project is to give network researchers an open simulation environment to
test their network protocols. After simulating a network using a network simulator
such as NS-3, a trace file is created which contains the events that occurred during the
simulation and based on these events, users study the network performance. There
are different types of files: trace file.tr and file. Pcap.
In this article, we have developed a new tool for automating the analysis of the
NS-3 trace file.tr in VANETs using multi-agent systems that automatically calculates
VANET performance parameters such as throughput, packet delivery rate, packet
loss, packets sent and packets received, it can also draw a graph. This tool calls Ns-3
Analyzer using Multi-Agents System (NsAMAS).
Our article is structured as follows Sect. 2 presents previous work. Section 3
presents the concept of agent and multi-agents system. Section 4 introduces
NsAMAS, presents its concepts and explains its architecture. In Sect. 5 we high-
light the most important and unique characteristics of NsAMAS. Section 6 explains
the uses and the obtained results. Finally, a conclusion and the future work are given
in Sect. 7.
2 Related Work
There are several analysis tools that have been developed to generate statistics on
network simulation performance metrics or parameters [7, 8]. For example, NS-2
Trace Analyzer JTrana, and Trace Graph. They use NS-2 trace files.
JTrana [9] is a NS-2 trace analyzer built from scratch in Java used to study wireless
network simulation using a graphical user interface. Its input is an NS-2 wireless
trace, and its output consists of many graphs or data sets of the network.
Trace Graph [10, 11] is another NS-2 trace file analyzer. This software offers
different analysis options. It is developed in MATLAB 6.0 [12]. This software also
allows the user to retrieve useful statistics from a specific NS-2 trace file.
NS-2 Trace Analyzer [13, 14] is a tool written in the C/C++ language. Similar
to the precedent tools. This tool does not generate graphs related to the recovered
statistical results.
A Novel Tool for Automating the Trace File Analysis ... 993
All the software mentioned above are designed for the analysis of the NS-2 trace
file. In addition, it does not allow the user to store the results of each trace file
analyzed in a database locally, so that they can reuse it without having to repeat the
analysis. To extract statistics from another previously analyzed trace file, the user
must download the trace file again. This task may be acceptable for a small trace file,
but this process can take a long time for simulations that generate large trace files.
In this article, we will try to resolve these issues.
The topic of multi-agents system (MAS), although not new, is currently a very active
research area. It is a specialty which is interested in collective behavior evoked by
the interaction between several autonomous and flexible entities called agents, these
interactions are manifested on cooperation, competition or coexistence between these
agents. We are going to present the key concepts needed by agents and multi-agents
system (MAS) to design a practical architecture for analyzing the NS-3 trace file.
3.1 Agent
The concept of agent is a whole concept born decades ago. In the literature, many
works have focused on agents and multi-agents system. Definitions differ from author
to author and from field of application to another. We can cite the most consensual
definitions that we have found [15] as follows:
– An agent is an autonomous entity or simply an independent entity, abstract or
real, able to act for himself and his environment and can communicate with other
agents in a multi-agents universe.
– An agent is an entity which presents the following characteristics: independence,
reactivity and anticipation and has a specific social capacity.
Multi-Agents System (MAS) [16] is a system made up of agents of the same type
or not, which cooperate and interact in a coordinated manner in order to achieve a
common goal or objective in an environment of which they have the same perception
and point of view.
A multi-agents system (or MAS) is a system that composed of the following
components [16]:
– An environment E, is a space which generally has a criterion.
994 S. Achchab et al.
– A set of objects O. These objects are localized. For each one it is possible to assign
a position in E, at a precise moment. These objects can be perceived, destroyed
and modified by agents.
– A set A of agents, is specific objects which contain the active entities of the system.
– A set of relations R that bring things together.
– A set of Op operations allow O objects to be interpreted, produced, consumed,
transformed and manipulated by agents.
4 NsAMAS Overview
NsAMAS is a new tool for automating the analysis of NS-3 trace file.tr in VANETS
using multi-agents system. The objective of NsAMAS is to perform a practical and
automatic analysis of the .tr trace file produced by a simulation using graphical
interfaces.
4.1 Architecture
This software is built on three layers: a presentation layer, a business layer and a
database access layer, as shown in Fig. 1.
• Presentation Layer
This layer presents a graphical interface which allows the user to interact with the
software. This layer contains the master agent to read the file, extract the fields
and transfer the data to another agent. User requests move from the presentation
layer to the actual processing layer. It also provides data in table and graph form
to study the results.
• Treatment Layer
This layer processes the data received from the source and stores the results in
the database. It contains two agents, one for calculating metrics and the other for
drawing graphs. The primary processor is responsible for analyzing the trace files
and storing the results, which is one of the basic software processes. It is also
responsible for studying general simulation information’s and each simulation
node. The Charts agent is responsible for plotting all types of graphics.
• Database Access Layer
This layer is used to store analysis results in the local database or to retrieve
requested data. It also interacts with the agent which calculates the performance
parameters to store the results, and with the master agent to retrieve the data
and store the information for each trace file.tr. The database used in NsAMAS is
MySQL and contains a table for each trace file entered by the user.
A Novel Tool for Automating the Trace File Analysis ... 995
One of the main objectives of NsAMAS is to facilitate the analysis of trace files.
Currently, NsAMAS focuses on NS-3.tr trace files.
The trace file in ASCII code contains 12 fields [2]. each file line begins with an
event (+, -, d, r) descriptor followed by the simulation time (in seconds) of that event,
and the from and to node. Then we have the type, the size of the packet (in bytes)
and Flags (ignore). The next field is the packet class, which can be used to identify
a particular TCP connection. The next two fields are the source address and the
destination address. Finally, the sequence number of the protocol and the identifier.
When a trace file.tr is sent as input, the agent reads the file, extracts the 12 fields
mentioned above for use in the processing phase, and finally stores the information
about this file in the database.
In this part, we identify the most important characteristics of NsAMAS and we also
present the architecture and the use of this software.
996 S. Achchab et al.
The first step of the analysis is to download the NS-3 simulation.tr trace file. Then the
file will be read and the information processing will start using an agent. The data on
the produced documents will be recorded in the database at the end of the processing
and analysis step. The general idea of the system, for analyzing, processing and
storing the file information, is shown in Fig. 2.
In this step our software starts to extract 12 fields from the trace file which were
provided as input. Once the trace file has been selected, an agent first tries to extract
the fields, then another agent uses these fields to process and analyze the data. Once
the processing phase of the trace file is completed, the results must be stored in the
database. Each trace file has its own table.
A Novel Tool for Automating the Trace File Analysis ... 997
In the trace file processing step, the NsAMAS can automatically calculate the general
simulation metrics for the.tr trace file using an agent and without any user intervention
and it calculates other information regarding the metrics for each node.
In this section, we explain in detail the different functions of NsAMAS and also
describe the information that NsAMAS can retrieve such as general simulation infor-
mation, general information about each node and graphics. The Table 1 provides the
simulation parameters that we used in our experiment.
To measure the network performance metrics, using our NsAMAS trace file analyzer.
An agent calculates various performance parameters: the general simulation param-
eters and the specific parameters to each node (Fig. 3). As an example of these
Table 1 Simulation
Parameters Values
parameters
Version of simulator NS-3.27
Routing protocol DSR
Number of vehicles 50
Simulation time 120 s
Physical and MAC layer 802.11p
Packet size 1000 bytes
Mobility model Manhattan Grid
998 S. Achchab et al.
metrics: throughput, packet drop rate, good throughput, packet delivery rate (PDR),
packet loss, packet loss rate (PLR), normalized routing load (NRL) and concentra-
tion activity. These metrics are the most common requested by researchers and the
most useful for evaluating the performance of VANETs. User can also retrieve the
results of previously scanned trace files from a local database as shown in Fig. 4.
At this point, the user can draw various graphs in the form of a curve or a histogram
as shown in Figs. 5 and 6. Once all parameters are set, graphics can be drawn for
general simulation or for a particular node. The graphics provided by this tool are
the packet delivery rate and the bit rate (bits/s) versus time. Finally, it is also possible
to export graphics as an image.
In this part, we provide synchronization results related to the basic steps described in
Sect. 5, which contains the reading of the trace file, analyzing and finally storing the
data in the local database. We have proved that the software not only handles trace
file.tr, but it does so extremely quickly. We present the results in Fig. 7 regarding the
above processes.
Based on the above results, it is clear that our NsAMAS software operates even
with large trace files to reduce the processing time. And we notice that the analysis
time increases with the size of the file.
We also provide a comparison of NsAMAS with other popular simulation analysis
tools. These tools are illustrated in Sect. 2 and are some of the best-known tools
dealing with simulated trace files. One of their main drawbacks is the time spent in
analyzing the trace file. The comparison shows that our NsAMAS software gives
good results in processing time Fig 8.
A Novel Tool for Automating the Trace File Analysis ... 1001
In this work, we presented a new tool labeled NsAMAS to automate the analysis of
VANETs simulation trace files, performed with NS-3 using the multi-agents system
concept. The main objectives of this tool are: offering a simple interface to facilitate
the processing of.tr trace files in a very short time for all file sizes. Furthermore,
displaying the results automatically in the form of tables and graphs to help students
and researchers studying the performance of VANETs under NS-3. Results can be
stored in the local database to reuse them without the re-analysis phase. Our NsAMAS
software provides good results in terms of processing time and analysis compared
to other popular simulation analysis tools.
Building on the success of this software, we intend to upgrade it to calculate other
metrics, such as transmission capacity and power consumption. We also intend to
use other big data tools to analyze the file and draw other types of diagrams.
References
Keywords Green optimization · Fuzzy logic · Base station and relay station
deployment · Energy consumption · On-off switching strategy
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1003
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_91
1004 S. Dahmani et al.
1 Introduction
In recent years, mobile data traffic is growing rapidly and yields very high produc-
tivity. Wireless networks have become a challenge for operators in current and future
years. That’s why researchers are trying to develop the network every time and
everywhere.
A basic requirement for green wireless communications is low energy consump-
tion and low air pollution. As the number of mobile stations has swelled and seen the
needs of users, mobile operators have to increase the number of Base Stations (BSs )
or Relay Stations (RSs ), which implies an increase of the energy consumer in the
wireless cellular network. According to [1], the annual number of BSs has enhanced
by two million for the period from 2007 to 2012, which conducts to uncontrolled
production of carbon dioxide.
The BSs used to cover remote areas are mainly operated by diesel generators. The
annual operating cost of these power grids is often very high.
As pointed by [2], in 2002, the production of CO2 related to information and
communication technologies (ICT) reached 151 metric tons of carbon dioxide
(MtCO2 ), nearly half of this value is due to the wireless connection. CO2 emis-
sions are also expected to attain 349 metric tons in 2020. The authors in [3] detect
that information and communication technology (ICT) consumes approximately 600
Terawatt-hour (TWh) of electrical energy or three percent of the total worldwide elec-
trical energy, this number is expected to grow to 1700 TWh in 2030. These statis-
tics convince researchers around the world to develop new ways to reduce energy
consumption and reduce its negative impacts on the environment.
In the telecommunication system, the RSs are distributed with BSs to increase
the connectivity rate of these networks as shown in Ref. [4, 5]. Therefore the mobile
stations (MSs) can send their data directly to the base station (BS) or by relay stations
(RSs) to relay their data to the BS as shown in Fig. 1. This technique is called a
multi-hop relay network.
The CO2 emissions and energy consumption by RS are lower than those by BS.
Moreover, the installation is easier and faster and the maintenance costs are lower.
Unlike BS, RS does not require a direct connection to the core network [6]. The
use of RS in a cellular network in parallel with the use of BS clearly has its own
unique characteristics. However, incorrect deployment in unsuitable locations can
lead to lower energy efficiency, throughput, transmission delay, transmission loss,
power consumption, and increased cost of deployment expenses. In order to properly
utilize the advantages of RS and BS, an effective strategy is required to choose the
method and how to deploy BS and RS to green cellular networks, to achieve the
desired target and throughput, with a reasonable deployment budget, less energy
wastage, and most of all, minimal environmental pollution.
Regarding the deployment of base stations in an environment, there is a set of
deployment strategies such as randomly, or at the edges of cells as indicated in
[7, 8].
A Fuzzy On/Off Switching Strategy … 1005
To install BS, the central position is chosen because it reduces the number of
stations to be installed and at the same time it covers a large number of mobile
stations with less disturbance [9]. Figure 2 shows an example of a base stations and
relay stations deployment site in area divided into grids.
1006 S. Dahmani et al.
Our first challenge, in this paper, is to propose an efficient placement scheme for
the deployment of BSs and RSs on a selected site taking into account the uncertainty
and imprecision of the coverage rate and traffic demand.
The second challenge is to minimize the power consumption of these stations by
using the on-off switching strategy while maintaining the network coverage as used
in our paper [10].
In Sect. 2 we describe the system model and problem formulation. Section 3
presents our proposed approach based on fuzzy logic and on-off switching strategy.
Section 4 shows the simulation Results and application. The conclusion is drawn in
the final Section.
Step 1:
We divide the area into divers’ grids, the publishing site is already chosen as a focal
point. The decision to install a BS or a RS in each site depends on the number of
users deployed in each network, as shown in Fig. 3.
Fig. 3 BS or RS deployment in an area divided into grids depending on the number of mobile
stations (MSs ) in each grid.
A Fuzzy On/Off Switching Strategy … 1007
In the literature, this deployment mainly depends on the threshold value of the
number of users which is calculated as follows:
T NU
Tr = (1)
T NG
where Tr is the threshold value, TNU is the total number of users and TNG is the
total number of grids.
The authors in [9] relied directly on this threshold to define the type of station
that will be installed. If the number of users (TNU) on each grid is higher than the
threshold value, BS is published. On the other hand, if this number is lower than the
threshold, the RS is displayed, as shown in Fig. 4.
In the existing model to define the number of users, the decision maker monitored
the traffic for a defined period (e.g. one month) and calculated the average value.
In our view, the average does not accurately reflect the user traffic. Therefore, we
consider that this value is an imprecise and uncertain value and not a constant one.
Step 2:
In the classic case, after the final publication of base stations and relay stations,
an installed station remains active all the time, even for periods when the traffic is
low, which leads to a loss of energy. The next step (challenge) of our work is the
implementation of the on-off switching approach for these stations to reduce energy
consumption (see [10]).
1008 S. Dahmani et al.
r = T N U
T (2)
T NG
where, T N U is a fuzzy number with the following membership function:
T NU −t
max0, 1 − α i f t ≤ T NU
μT (t)
NU max 0, 1 − t−T NU
β
if t > T NU
where the two values α and β are positive fixed values signifying respectively the
left and right deviations of the fuzzy numbers. The membership function μT NU (t)
is represented in Fig. 5.
Fig. 5 Illustration that represents the membership function of the fuzzy number T NU.
A Fuzzy On/Off Switching Strategy … 1009
Each BS or RS station has two main powers, the basic and transmitting energy. The
basic energy is small energy related to these stations when they are in an idle state,
while the transmission power is related to them when sending information to the
users and thus it is in an active state, this means that the power consumption can be
reduced by controlling the BS or RS mode. When BS or RS is not fully used or not
connected to the maximum number of users, it can be kept in standby or idle mode
to minimize power consumption.
We will rely on the same strategy used in our previous article [8], which is based
on the concept of traffic states, the reactive approach and the proactive approach.
Traffic States: Suppose that in a given area and for one day, the traffic is distributed
as follows (see Fig. 6).
Between 00 am and 10 am, the traffic does not exceed 450 GBytes;
Between 10 am and 2 pm, the traffic is between 450 and 900 GBytes;
Between 2 pm and 8 pm, the traffic exceeds 900 GBytes;
Between 8 pm and 00 am, the traffic is between 450 and 900 GBytes;
The maximum traffic during the day is 1200 GBytes.
From this data, traffic variations can be divided into three different states s1 , s2
and s3 where s1 = [0; 450], s2 = [450; 900], and s3 = [900; 1200].
Reactive Approach: We consider the highest traffic state. Next, we determine the
optimal set of BS for that state. As the traffic conditions decrease, we need to
deactivate a number of BS while still meeting the coverage requirements.
In order to carry out our simulation study, we have considered a large network area
to be covered. The size of this area is 10 × 10 km2 divided into 40 grids. The number
of mobile stations (MS) to be covered is 1200. Fixed power consumption (P) at the
BS and RS is set, respectively, to 68.8 W and 20 W (see [11] for instance). For
traffic reports, we randomly generated this traffic over a day spanning one month
and we calculated the average traffic per each hour. We obtained three traffic states
as follows:
Base stations and relay stations consume the majority of energy in the cellular
network. Therefore, activating and deactivating these components will lead to signif-
icant energy savings and reduced CO2 emissions. Electrical energy is derived from
fuel oil, where each one kilowatt hour represents 620 g of carbon dioxide (see [12]).
The expression of carbon dioxide emissions for each traffic state is defined by:
where,
CO2 (si ): The amount of CO2 produced by the traffic state si ;
Nsi : The number of BSs and RSs started from the traffic state si ;
PBS (W): The energy consumption of the BS;
PRS (W): The energy consumption of the RS;
Tsi (H): The duration of the traffic state.
In the classic approach, the decision makers decide to keep all the 40 BSs active
throughout the day, and then the total CO2 emission, in KgCO2 /Day, is
40 × 68.8 × 24 × 620
= 40.94976
106
When we use the switching on-off strategy considering only the BSs, the CO2
emissions for each traffic state is (Fig. 6).
40 × 68.8 × 6 × 620
C O 2 (s3 ) = = 10.2374
106
26 × 68.8 × 8 × 620
C O 2 (s2 ) = = 8.87244
106
16 × 68.8 × 10 × 620
C O 2 (s1 ) = = 6.82496
106
In our approach, after the fuzzy installation of relay stations in parallel with base
stations and the application of the reactive approach during the day, the total CO2
emission is:
By comparing these results, we see that the CO2 emission is reduced by 17%,
which shows the efficiency of our approach.
1012 S. Dahmani et al.
5 Conclusion
References
1. Mclaughlin S, Grant PM, Thompson JS, Haas H, Laurenson DI, Khirallah C, Hou Y, Wang
R (2011) Techniques for improving cellular radio base station energy efficiency. IEEE Wirel
Commun 18:10–17
2. Alsharif MH, Nordin R, Ismail M (2013) Survey of green radio communications networks:
techniques and recent advances. J Comput Netw Commun
3. Humar I, Ge X, Xiang L, Jo M, Chen M, Zhang J (2011) Rethinking energy efficiency models
of cellular networks with embodied energy. IEEE Netw 25:40–49
4. Yang Y, Hu H, Xu J, Mao G (2009) Relay technologies for WiMAX and LTE-advanced mobile
systems. IEEE Commun Mag 47:100–105
5. Chen L, Huang Y, Xie F, Gao Y, Chu L, He H, Li Y, Liang F, Yuan Y (2013) Mobile relay in
LTE-advanced systems. IEEE Commun Mag 51:144–151
6. Arthi M, Arulmozhivarman P (2017) Power-aware fuzzy based joint base station and relay
station deployment scheme for green radio communication. Sustain Comput Informatics Syst
13:1–14
7. Son K, Oh E, Krishnamachari B (2011) Energy-aware hierarchical cell configuration: from
deployment to operation. In: 2011 IEEE conference on computer communications workshops
(infocom wkshps), pp 289–294 (2011)
8. Mohammed G, Soufiane D, El Bekkaye M, Abdelhafid S (2019) Optimization of multi-
objective and green LTE RNP problem. In: 2019 International Conference on Wireless Tech-
nologies, Embedded and Intelligent Systems, WITS 2019 (2019) https://doi.org/10.1109/
WITS.2019.8723811.
9. Ameer R, Rolly RM, Jacob J (2018) Dynamic base station or relay station deployment and
small cell On/Off strategy for green ultra dense networks. Procedia Comput Sci 143:827–834
10. Dahmani S, Gabli M, Mermri EB, Serghini A (2020) Optimization of green RNP problem for
LTE networks using possibility theory. Neural Comput Appl 32. https://doi.org/10.1007/s00
521-018-3943-x
11. Ratheesh R, Vetrivelan P (2019) Energy efficiency based on relay station deployment and sleep
mode activation of eNBs for 4G LTE-A network. Automatika 60:322–331
12. Koutitas G (2010) Low carbon network planning. In: 2010 European Wireless Conference
(EW), pp 411–417
Multi-band Planar-Inverted-F-Antenna
Design for WIFI WIMAX and WLAN
Applications
Asma Khabba, Layla Wakrim, Saida Ibnyaich, and Moha M’Rabet Hassani
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1013
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_92
1014 A. Khabba et al.
The multiband PIFA antenna has attracted great interest for the employment in
wireless devices, in addition its miniature dimensions are compatible with mobile
terminals and warrants versatile functions of a compact mobile handset [5]. Several
techniques are used to achieve multi-band performance such as slot insertion [6–8],
the addition of a parasitic element [9], the addition of a capacitive load [10].
In this manuscript, a novel triple-band PIFA antenna using L-shaped slots is
designed and studied using CST Microwave Studio software. The Proposed antenna
is designed in such manner to accommodate the three applications WIFI, WIMAX
and WLAN in single antenna. The antenna is marked with high gain at three resonant
frequencies. While the antenna evolution process is carried out to investigate the
effect of the etched slots on the resonant frequencies. In addition, the simulated
results concerning current distribution and radiation patterns are discussed.
The rest of this paper is organized as follows: Sect. 2 presents the antenna structure
and geometry. The antenna operating mechanism is discussed in Sect. 3. While Sect. 4
describes the simulating results including reflection coefficient, input impedance and
distribution current. The gain and radiation patterns are presented in Sect. 5. Finally,
Sect. 6 concludes the study.
The antenna geometry is presented in Fig. 1. The suggested design consists of a radi-
ating plate which is connected to the ground plane by a short circuit of width 2.5 mm
and powered with a plate of width 4.363 mm. The ground plane is made up using the
low cost FR4 epoxy substrate with dielectric constant of 4.3, loss tangent of tanδ =
0.018 and thickness of 1.2 mm. The dimension of the radiating plate, ground plane,
shorting plate and the powered plate is displayed in Fig. 1(a). While the dimensions
of the L-shaped slots is showed in Table 1.
Figure 2 presents the stepwise geometries through which the proposed design has
been obtained. While the reflection coefficient of the antennas at the four stages is
presented in Fig. 3. As can be seen the design is started with a conventional PIFA
antenna which exhibits a single band operation where the resonant frequency is
obtained at 4.75 GHz with 850 MHz of bandwidth. The second step is came out
with the target to create a novel operating band. The second antenna in Fig. 2(b) is
obtained by inserting an L-shaped slot of size 19.365 × 5 mm2 in the radiating plate.
As can be seen in Fig. 3 the loaded slot leads to create a novel resonant frequency at
2.8 GHz, however the new resonant mode does not have an appropriate impedance
matching (S11 = −8 dB). The third step is performed by inserting a second L-shaped
slot of size 22.8 × 5 mm2 in the radiating plate (Fig. 2(c)). As remarked from Fig. 3,
Multi-band Planar-Inverted-F-Antenna Design … 1015
the effect of the second slot is presented in the creation of the third resonance at 2.3
GHz while the impedance matching does not reach an appropriate value. Finally, the
last step is done with the cutting of another L-shaped slot of size 9.457 × 8 mm2 .
It is clearly remarked that this cutting slot has a crucial role to adjust and improve
the impedance matching of the three resonant modes. Which makes it ready to be
employed for WIFI, WIMAX and WLAN standards. The details about the three
resonant modes is described in the next subsection.
1016 A. Khabba et al.
Fig. 2 Configuration of the proposed PIFA antenna a conventional antenna, b antenna 2 c Antenna
3, d proposed antenna
4 Simulation Results
is obtained at 2.47 GHz (S11 = −16.08 dB) which corresponds to WIFI standard,
with 120 MHz of bandwidth from 2.41 GHz to 2.53 GHz. While, the second resonant
frequency is obtained at 3.51 GHz (S11 = −21.77 dB) which corresponds to WIMAX
standard with 40 MHz of bandwidth from 3.49 GHz to 3.53 GHz and the third
resonant frequency is obtained at 5.23 GHz (S11 = −32.11 dB) which corresponds
to WLAN standard with 390 MHz of bandwidth from 5.04 GHz to 5.43 GHz.
The input impedance of an antenna is expressed as Z in = Rin + j X in , where
Z in is the input impedance, Rin is the resistance (real part) and X in is the reactance
(imaginary part). In order to obtain a good impedance matching at the resonant
frequency the input impedance should be close to 50 which means the input
resistance Rin should be close to 50 and the input reactance X in should be close
to 0. The input impedance of the proposed PIFA antenna is presented in Fig. 5.
Fig. 6 Current distribution at the resonant frequencies, a 2.47 GHz, b 3.51 GHz, c 5.23 GHz
As can be seen, the input resistance at the three resonant frequency is Rin is 39.15
, 54.74 and 48.01 respectively and the input reactance X in is −8.96 , 7.12
and −1.4 respectively. It should be noted that the best impedance matching is
reached at the third resonance frequency where Rin is close to 50 and X in is close
to 0 which generates a low reflection coefficient of −32.11 dB.
To better understand the antenna behavior, the distribution of current at the three
resonant frequencies is depicted in Fig. 6. As depicted in Fig. 6(a) at the first resonant
frequency the maximum of current is concentrated at the horizontal part of the second
slot and a slight current intensity is concentrated in the horizontal part of the third
slots. While at the second and third resonant frequencies (Fig. 6b, c), it is remarked
that the maximum of current is intensified around the vertical and horizontal parts
of the second and third slots). Therefore, it can be concluded that the second and the
third slots are responsible to generate the resonant frequencies while the first slot
is responsible to enhance the impedance matching as well as tuning the resonance
frequencies values.
The 3D radiation patterns at 2.47 GHz, 3.51 GHz and 5.23GHz are presented
respectively in Fig. 7(a), (b) and (c), while the 2D radiation patterns at the two cutting
plan phi = 0° and phi = 90° are presented in Fig. 7(d), (e) and (f). As can be seen
the proposed antenna provides high radiation performance in term of gain where the
max gain value obtained at 2.47 GHz, 3.51 GHz and 5.23 GHz are respectively 4.09
dB, 5.45 dB and 8.5 dB while the maximum radiation is oriented toward broadside
direction (phi = 0°, theta = 0°) for the first and third resonant frequencies and toward
the positive part of x axis for the second resonance.
Multi-band Planar-Inverted-F-Antenna Design … 1019
5 Radiation Patterns
6 Conclusion
Fig. 7 Current distribution at the resonant frequencies, a 2.47 GHz, b 3.51 GHz, c 5.23 GHz
1020 A. Khabba et al.
References
1 Introduction
Vehicular Ad hoc Networks (VANETs) are a special case of Mobile Ad hoc Networks
(MANETs), where the mobile nodes are intelligent vehicles. VANETs have particular
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1021
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_93
1022 S. Ajjaj et al.
characteristics such as the high mobility, the dynamic topology, the delay constraints
and the intermittent connectivity [1]. Thus, the routing in VANETs must be as effi-
cient and reliable as possible. In this context, the present work introduces a new
approach for studying the effect of factors on the performance of the routing process
in VANETs. Our proposed method will offer the possibility to simultaneously assess
the effect of numerous factors with fewer experiences, by applying the Plackett-
Burman method [2]. The aim of our new approach is to model mathematically the
effect of factors on variables called responses. To perform our model, we conduct
eight experiments integrating four independent factors namely the Black Hole Attack,
the Worm Hole Attack, the Node Density and the Number of Connections. Accord-
ingly, we select Throughput, Packet Loss Ratio, Average End to End delay and
Routing Overhead as response variables. To implement our work, experiences are
built using Network Simulator NS-3, SUMO and AODV routing protocol [3].
The remainder of this paper is structured as follows. In Sect. 2, we present AODV
routing protocol. In Sect. 3, we will discuss the related work. Section 4 then introduces
the Plackett-Burman method. Section 5 will deal with the details of our proposed
work. Analysis of the acquired results is detailed in Sect. 6. Finally, we conclude
with synthesis of the work achieved and future research directions.
2 Overview of AODV
3 Related Work
Multiple researches have dealt with the performance analysis of AODV in VANETs.
For instance, the authors in [4] evaluated the performance of three VANET routing
protocols, AODV, DSR and OLSR under two scenarios: urban and highway, in terms
of Packet Delivery Ratio and End to End delay in different node density. They
conclude that AODV has better performances regarding to packet delivery ratio in
both scenarios, while DSR has better delay, whereas OLSR showed average perfor-
mances in both scenarios. In [5], the authors studied the effect of node density on
Throughput, Packet Delivery Ratio (PDR), End to End delay and Packet Loss Ratio
using DSDV, OLSR and AODV. The comparison shows that AODV gives better
results. [6], compared throughput and delay of AODV and DSR. The results show
that AODV has better Throughput than DSR but more delay. In [7], the researchers
simulated a VANET using NS2 and VanetMobiSim. They compared three routing
protocols DSDV, AOMDV and AODV in terms of Packet Delivery Ratio and End to
End Delay. They stated that reactive routing protocols provide good results in low
and middle node densities. In [8] a modified protocol of AODV for multi-path was
proposed to overcome the problem of AODV high routing overhead. The authors
of paper [9] studied the effect of Black Hole attack on the performance of AODV
and ZRP, in terms of average Throughput, Packet Delivery Ratio, End-to-End Delay,
Normalized Routing Load and Average Path Length. The findings show that the
performance of AODV is decreased when Black Hole attack is launched. [10], illus-
trates the vulnerability of AODV to Black Hole Attack in VANETs and propose a
new algorithm to secure it. In this paper [11], authors evaluated the impact of Worm
Hole Attack on AODV, in terms of Average Number of Hopes per Route, Average
Delay, Average Route Discovery time, Average Throughput, Average Retransmis-
sion Rate, Average Data Dropped, and Average Traffic received. They emphasize the
degradation of AODV under Worm Hole Attack.
2. Build the PB matrix of experiments: the table where columns represent the
Factors and rows indicate the number of experiments. The matrix contains only
+1 and −1.
3. Perform the matrix of effects denoted X. It is based on the matrices of
HADAMARD [12]. We build it by adding a column containing only +1 to
the left of the matrix of experiments. This matrix is then used to calculate the
coefficients of the mathematical model. X is such that:
XT X = nIn (1)
ŷ = b0 + b1 x1 + b2 x2 + b3 x3 + · · · + bk xk (2)
Y = BX (3)
N
N
EA = 2 × [ (yi) − (yi)]/N (5)
i =1 i =1
A = +1 A = −1
Where N is the number of experiments and y represent the values of the response.
A New Approach for Evaluating the Performance of AODV Routing Protocol … 1025
5 Proposed Work
One of the main constrictions in VANETs is security wherein the vulnerability mainly
originates from the wireless communication and the lack of infrastructure [1]. There-
fore, various types of attacks can be launched in the network, e.g., the Black hole
Attack and the Wormhole Attack. In the Black hole Attack, the malicious vehicle
falsely pretends to have a short and fresh route to the destination and sends a fake
route reply with a high sequence number to the source node. The source node begins
the data transmission, believing that the optimal path is through this malicious node.
The malicious node then captures all routing packets and drops them. Worm Hole
Attack is carried out in a cooperative way, where two colluding nodes create a virtual
tunnel between them (at remote location) and let other nodes think that the routing
is faster via them. Therefore, fake routes are created resulting in a virtual traffic in
the network. The goal is to change the network topology and confuse the routing
process.
Other constraints influence the routing in VANETs. The findings in [7] and [14]
show that the routing performances vary considerably in terms of density and the
number of connections.
In our study we choose to study the effect of four factors: the Black Hole Attack,
the Worm Hole Attack, the Node Density and the Number of Connections.
For each factor, we define the levels low (−1) and high (+1) (Table 1).
For Black Hole and Worm Hole attacks, we use logical values (0 for level −1 and
1 for level +1). Node Density varies from 50 to 150 vehicles, whereas Number of
Connections varies from 4% of the total number of vehicles in level −1, and 20% in
level +1.
In our study, we select Throughput (y1), Packet Loss Ratio (y2), Average End to
End Delay (y3) and Routing Overhead (y4) as response variables.
1. Throughput (Th) is the total number of bits successfully received during
simulation time, it is measured in kilobits per second (kbps).
2. Packet Loss Ratio (PLR) represents the ratio of the number of lost packets to
the total number of sent packets.
3. End to End Delay (E2E delay) is the average time needed to transmit data from
source to destination, it is measured in (ms).
4. Routing overhead (RO) is the total number of control packets generated by
AODV routing protocol during route discovery and route maintenance.
Table 1 The values for levels (+) and (–) of all factors
N Factor High level (+1) Low Level (−1) Factor code
1 Black hole attack 1 0 X1
2 Wormhole attack 1 0 X2
3 Node density 150 50 X3
4 Number of connections 20% 4% X4
1026 S. Ajjaj et al.
In this study, A simulation map of El Jadida city Morocco is imported from Open
Street Map (OSM). Simulation parameters are described in (Table 2).
6 Experimental Results
The coefficients of the mathematical model are calculated by applying the method
of least squares. The calculations are done using a spreadsheet. The resulting model
for each Response ŷ is presented below:
where: x1, x2, x3 and x4 are respectively the factors of Black Hole Attack (BKH),
Worm Hole Attack (WMH), Node Density (ND), and Number of Connections (NC).
Plots of Effects
Figure 1 illustrates the effect of all Factors on the AODV throughput. We notice that
the throughput decreases in + l level of all Factors. For Worm Hole and Black Hole
Attacks, this decrease is due to the presence of malicious nodes that try either to drop
packets in case of Black Hole Attack or capture and tunnel the packets to a distant
location in case of Worm Hole Attack. As a result, the number of successfully received
packets reduces causing the degradation of the throughput. AODV Throughput also
cuts down with the increase in Node Density and Number of Connections.
According to Placket Burman Method, we can rank these factors based on the
average effects. The effect of Worm Hole Attack comes first, then the Black Hole
Attack, the Node Density and finally the Number of Connections. The results
achieved using PBM agree with several works in literature. For example, the authors
in [5] reported that the throughput decreases under the effect of Node Density. The
same conclusion is achieved in [14] regarding to the Number of Connections. Further-
more, the study carried out in [10] confirmed that AODV throughput is better in
scenarios free of routing security attacks.
Examining Fig. 2 related to the effects on AODV Packet Loss Ratio. It is clear
that the four Factors have a positive effect in their +1 level. Meaning that, the Packet
Loss Ratio increases in the presence of Black Hole and Worm Hole Attacks as well
as when the network is dense or the Number of Connections is high. According
to PBM, the most influencing factors are respectively Worm Hole and Black Hole
Attacks. Node Density comes in third position and Number of Connections in the
last rank.
The outcomes revealed by Plackett-Burman Method are in line with other
researcher’s study. For instance, the results in paper [5] emphasize that the number
of lost packets grows proportionally with the number of nodes. Additionally, [15]
stated that the Worm Hole Attack is responsible for rising AODV packet loss. In [9]
the conclusion is that AODV performance significantly degrades under Black Hole.
Figure 3 indicates the effect on End To End delay. We notice that the effect of
Black Hole Attack is negative unlike the three other factors with positive effect. In
other words, the End To End delay decreases in the level +1 of Black Hole Attack,
while it increases in the level +1 of the other factors. This rise is very important for
Number of Connections, while it slightly grows for Worm Hole Attack and Node
Density.
The reason is that AODV uses the mechanism of route discovery before sending
data packets. As the number of connections increases, the route search time will also
increase leading to high values of delay. Additionally, in Worm Hole Attack, packets
are tunneled through the malicious nodes that are located at remote positions; the
End To End Delay then becomes higher.
The results of PBM show that AODV performance decrease in the presence of
Worm Hole Attack, also in high dense networks and increased Number of Connec-
tions. The exception found is that the End To End Delay for Black Hole Attack is
reduced when the attack is launched. This is explained by the behavior of Black Hole
node that sends fake and fast Route Reply (RREP) to source node, this significantly
decreases both the route discovery process and the delay. This conclusion is compat-
ible with [9] where it is concluded that delay under Black Hole is better than in a
normal scenario. [4] Confirms the increase of AODV End To End delay with the rise
in the number of nodes and number of connections.
The authors in [16] studied the Worm Hole delay and proposed an improved
AODV to preserve the network from this attack.
The results of the effect on Routing Overhead are shown in Fig. 4. We observe
that the Routing Overhead increases in +1 level of all the Factors except for Black
Hole Attack that shows insignificant effect. This result copes with the study in [17],
which proved that the Routing Overhead decreases under Black Hole Attack because
malicious node sends a fake RREP as fast as possible. Consequently the Overhead
decreases because the route discovery packets are reduced.
We also note that Node Density has the most significant effect, followed by
Worm Hole Attack and finally the Number of Connections. This rise in Overhead is
caused by the large number of Route Request messages generated by AODV, mainly
in VANET environment qualified by high mobility and frequent connections and
disconnections of nodes.
The outcomes of our study match with several previous studies. [8] stated that
the Routing Overhead of On-Demand routing protocols rises significantly when
the number of nodes increases. In [18] the researchers emphasized the importance
of implementing new mechanisms to reduce AODV control messages overhead.
The authors developed two improved versions of AODV protocol (AODV-LAR and
AODVLine) to overcome the issue of high routing over-head. The authors in [16]
confirm the negative impact of Worm Hole Attack on AODV overhead. They also
presented a new approach to reduce AODV Overhead, and secure it against Worm
Hole Attack.
Fig. 5 Experiment Vs
Estimated throughput
Fig. 6 Experiment Vs
Estimated PLR
Fig. 7 Experiment Vs
Estimated E2E delay
Fig. 8 Experiment Vs
Estimated overhead
A New Approach for Evaluating the Performance of AODV Routing Protocol … 1031
7 Conclusion
In this paper, we presented a new approach for the performance analysis of routing
protocols in Vehicular Ad hoc networks. As per our humble knowledge, our proposed
work unprecedentedly applied the Plackett-Burman Method, in which the findings
eventually cope with previous studies in the same area. This method allows to simul-
taneously evaluate the effect of variables called factors on other variables called
responses through a few number of experiences. Moreover we performed a mathemat-
ical model that describes the relationship between each response and the coefficients
of the factors.
In this study, we investigated the effect of four factors namely the Black Hole
Attack, the Worm Hole Attack, the Node Density and the Number of Connections on
the following performance metrics of AODV routing protocol: Throughput, Packet
Loss Ratio, End To End Delay and Routing Overhead. The achieved results demon-
strate the degradation of AODV performances when the attacks are launched in the
networks, as well as when the network is dense with high number of communica-
tions. The present work can be extended to appraise the effect of more factors on the
performance of other VANET routing protocols.
References
1. Arif M, Wang G, Zakirul Alam Bhuiyan M, Wang T, Chen J (2019) A survey on security attacks
in VANETs: Communication, applications and challenges. Veh Commun 19:100179
2. Plackett RL, Burman JP (1946) The design of optimum multifactorial experiments. Biometrika
33(4):305–325. https://doi.org/10.1093/biomet/33.4.305lastaccessed2020/10/26
3. Perkins C, Belding-Royer E, Das S (2003) Ad hoc On-Demand Distance Vector (AODV)
Routing. RFC 3561, RFC Editor https://www.rfc-editor.org/rfc/rfc3561.txt. Accessed 26 Oct
2020
4. Pranav K, Kapang L, Tuithung T (2011) Simulation based analysis of Adhoc routing protocol
in urban and highway scenario of VANET. Int J Comput Appl 12(10):42–49
5. Gupta P, Chaba Y (2014) Performance analysis of routing protocols in vehicular ad hoc
networks for CBR applications over UDP connections. Int J Eng Comput Sci 3(6)
6. Ghori MR, Sadiq AS, Ghani A (2018) VANET routing protocols: review, implementation and
analysis. J Phys Conf Ser 1049:012064
7. Vidhale B, Dorle SS (2011) Performance analysis of routing protocols in realistic environment
for vehicular ad hoc networks. In: 2011 21st International Conference on Systems Engineering,
pp 267–272
8. Hu S, Jia Y, She C (2017) Performance analysis of VANET routing protocols and implemen-
tation of a VANET terminal. In: 2017 International Conference on Computer Technology,
Electronics and Communication (ICCTEC), pp 1248–1252
9. Purohit KC, Dimri SC, Jasola S (2017) Mitigation and performance analysis of routing proto-
cols under black-hole attack in vehicular ad-hoc network (VANET). Wireless PersCommun
97:5099–5114
10. Tyagi P, Dembla D (2017) Performance analysis and implementation of proposed mechanism
for detection and prevention of security attacks in routing protocols of vehicular ad-hoc network
(VANET). Egyptian Inf J 18:133–139
1032 S. Ajjaj et al.
11. Ojha M, Kushwah RS (2014) Impact and performance analysis of WORMHOLE attack on
AODV in MANET using NS2. Int J Sci Res (IJSR) 3(6):1823–1826
12. Kharaghani H, Tayfeh-rezaie B (2005) A Hadamard matrix of order 428. J Comb Des 13
13. Dodge Y, Rousson V (1993) Alternative Methods of Regression. A Wiley-Interscience
Publication
14. Spaho E, Ikeda M, Barolli L, Xhafa F (2013) Performance comparison of OLSR and AODV
protocols in a VANET crossroad scenario. In Park JJ, Jong H, Barolli L, Xhafa F, Jeong HY
(eds) Information Technology Convergence, Lecture Notes in Electrical Engineering. Springer
Netherlands, Dordrecht, pp 37–45 (2013)
15. Narayanan SS, Murugaboopathi G (2018) Modified secure AODV protocol to prevent
wormhole attack in MANET. Concurr Comput Pract Exp 32
16. Safi SM, Movaghar A, Mohammadizadeh M (2009) A novel approach for avoiding wormhole
attacks in VANET. In: 2009 First Asian Himalayas International Conference on Internet, pp
1–6
17. Chavan AA, Kurule DS, Dere PU (2016) Performance Analysis of AODV and DSDV Routing
Protocol in MANET and Modifications in AODV against Black Hole Attack. Procedia Comput
Sci 79:835–844. Proceedings of International Conference on Communication, Computing and
Virtualization (ICCCV)
18. Ayash M, Mikki M, Yim K (2012) Improved AODV routing protocol to cope with high overhead
in high mobility MANETs. In 2012 Sixth International Conference on Innovative Mobile and
Internet Services in Ubiquitous Computing, pp 244–251
Design and Simulation of a New Dual
Band Fractal Antenna for TM
Applications
Abstract In this paper, we present the three first iterations design of a new dual-band
antenna based on U fractal geometry feeding by microstrip inset feed. The proposed
antenna is printed on FR4 substrate with a dielectric constant of 4.4. At the third
iteration, the studied antenna has a dual-band behavior with tow resonant frequencies:
2.25 and 3.29 GHz with a good impedance matching. The simulated results performed
by CADFEKO a Method of Moments (MoM) based Solver. The antenna offers low
profile, low coast and a very light weight which satisfy the requirement of MicroSat
Telemetry (TM) applications.
1 Introduction
Interest in multi-band antennas is growing, in particular with the aim of reducing the
number of onboard and ground antennas and by combining several applications on the
same antenna. In particular, the increasing diffusion of radio navigation systems such
as GPS (Global Positioning System) has made it possible to flood the world market
with a large number of new receivers combining good performance, small size and
reduced consumption [1–3]. In addition, the deployment of the new radio navigation
system European Galileo (planned for 2013) and its combination with GPS will
allow the implementation of more reliable, high-precision applications. Thus, the
development of new antenna architectures combining two or more radio navigation
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1033
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_94
1034 A. Moutaouakil et al.
2 Antenna Design
ln(N ) ln(5)
d= = = 1.464
ln 1 R ln(3)
3 Simulation Results
Fig. 3 Simulated reflected coefficient (S11) versus frequencies for three first iterations
Fig. 4 Simulated radiation pattern in E-plan and H-plan at: a 2.25 GHz b 3.29 GHz
4 Conclusions
In this work, the setup of U-shape fractal on the patch antenna allows a behavior
multiband. The simulated results performed by CADFEKO a MoM based Solver are
encouraging. We see that with the development of this antenna in terms of form, it
can be used for the MicroSat Telemetry (TM) application, which is characterized
by a resonance frequency equal to 2.245 GHz, and a bandwidth wide [2200–2290]
MHz.
References
1. Kumar D, Singh RB, Kaur R (2019) Global positioning system. In: Spatial Information
Technology for Sustainable Development Goals. Springer, Cham
2. Hadj Abderrahmane L, Brahimi A A new design of dual band fractal antenna for LEO applica-
tions. The Eighth International Conference on Systems and Networks Communications. ISBN
978-1-61208-305-6
3. Wibisono G, Wildan M, Wahyudi J, et al (2020) Co-design structure of dual-band LNA and
dual-band BPF for radio navigation aid application. Wireless Pers Commun
4. Hebib S, Aubert H, Pascal O, Fonseca N, Marc Lopez J (2007) Antennes multi-bandes pour
application GPS/Galileo/TéléMesure MicroSat. 15èmes Journées Natio ales Microondes, May
2007, Toulouse, France, pp 1–4
5. Mandelbrot BB (1983) The Fractal Geometry of Nature, San Francisco, pp 152–180
6. Constantine Balanis A (1997) Antenna Theory Analysis and Design. Wiley, Hoboken
7. Moutaouakil A, Jabrane Y, Reha A, Koumina MA (2019) Effect study of a ring on the
performances of a patch antenna. In 2019 International Conference of Computer Science and
Renewable Energies (ICCSRE), Agadir, Morocco, pp 1–3
8. Chen WL, Wang GM (2008) Small size edge-fed Sierpinski carpet microstrip patch antenna.
PIERS C, vol. 3
9. Gupta S, Kshirsagar P, Mukherjee B (2018) Sierpinski fractal inspired inverted pyramidal DRA
for wide band applications. Electromagn 38(2):103–112
10. Moutaouakil A, Jabrane Y, Reha A, Koumina MA (2020) Design and simulation of a fractal
multiband antenna printed on a circular patch. Int J Adv Sci Technol 29(06), 5290–5296
11. Boutejdar A, Halim BI (2019) Design and manufacturing of a dual patch ring antenna using
parasitic ring resonators and partial ground plane for multiband applications. In: 2019 IEEE
International Symposium on Phased Array Systems and Technology
Design and Simulation of Three Bands
Planar Inverted F Antenna Array for 5G
Communication Systems
Abstract The fifth generation of wireless mobile communication is one of the most
recent technologies in the field of wireless communications; this advance should
provide a very great service in terms of speed and diversity.
The old phones had an exterior antennas as an antenna dipole, but now most of
them had an integrated antennas, these antennas are PIFA (Patch Inverted F antenna),
a patch short—circuited in a corner (to decrease its size) and a power supply beside
it.
This paper is focused on this type of antenna that is frequently used in the field
of mobile communication, and it aims to present a design and simulation of a three-
band planar inverted F antenna array, which will be used as equipment under test for
electromagnetic compatibility tests in the fifth generation context.
1 Introduction
S. Arsalane (B)
Labo RITM, University Hassan II, Casablanca, Morocco
N. Arsalane
Labo LREA, Institut Superieur de genie applique, Casablanca, Morocco
M. Rifi
Labo RITM, University Hassan II, Casablanca, Morocco
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1039
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_95
1040 S. Arsalane et al.
a few kilometers. either, the installation of such networks does not need a heavy
upgrading of existing infrastructure as is the case of wired networks (digging trenches
to route cables, trucking and connectors…), which guide to a rapid development
of this type of technologies. On the other hand, there is a problem of regulation
relating to radio transmissions. Indeed, radio electric transmissions save for a large
applications (military, amateurs…), however they are sensitive to interference, that’s
why regulation is necessary in each country in order to define the frequency ranges
and the powers at which it is possible to transmit for each category of use [2].
Mobile telephony or cellular telephony is a way of telecommunication more accu-
rately of radio communication. The transmission of voice and data using radio waves
(frequencies in the UHF bands ranging from 700 to 2600 MHz) between a base
stations that can cover an area of several tens of kilometers in radius and the user’s
mobile phone [3].
Antennas are a cornerstone of wireless telecommunications systems, these
elements allows accessing radio, television telephony wherever they are, and they
are present at both ends of a link. Wireless transmission cannot be performed without
the use of transmitting and receiving antennas. On transmission, it is the only inter-
face capable of ensuring signal transfer between a guided propagation medium and
free space. At the reception, it takes care of the interfacing between the free space
and the guided environment [4]. Among the most used antennas, it is PIFA antenna
(inverted F antenna) which is increasingly used in the mobile phone market, and it
is resonant at a quarter wavelength (thus reducing the required space needed on the
phone) [5]. This paper aims to study the PIFA antenna characteristics and simulation
of a three band PIFA antenna array for 5G communications systems that will be used
as equipment under test for electromagnetic compatibility tests in incoming paper.
This article is organized in three parts:
• Presentation of PIFA antenna and its characteristics.
• Implementation of a model of the PIFA antenna and PIFA array antenna.
• Results and discussion.
In Recent years, the development of mobile telephony has growing rapidly, new
bands have been added, and the market is demanding small mobile phones with
more services, which requires the use of miniature antennas, multiband and which
minimize the risk of parasitic radiation to the human body [6].
PIFA antennas (Planar Inverted F antennas) are today the most used in mobile
telephony, due to their simple and flexible structure: reduced volume, low manu-
facturing cost, facility of integration; however, their major disadvantage is having a
narrow bandwidth [7].
Design and Simulation of Three Bands Planar Inverted F Antenna Array … 1041
Fig. 1 Typical inverted F antenna (left) and typical PIFA antenna (right)
The PIFA antenna is the consequence of the modification of the IFA antenna (inverted
F antenna) from a horizontal wire element to a planar structure in order to rectify its
losses mismatch and improve its radiation characteristics [8] (Fig. 1).
This structure consists of a metallic radiating element parallel to the plane of
mass. One of the edges of the patch is connected to the ground by a short circuit
plane which constitutes a particularities of this antenna. The upper element is usually
excited by a probe coaxial that penetrates the dielectric substrate [9].
The variation of its dimensions, the power supply position, the plane of short
circuit, the substrate height alter the performance of PIFA.
These antennas have a several advantages like:
• Their compactness, low manufacturing cost.
• Their performance can be improved by adding loads judiciously.
• In research carried out on mobile telephony: PIFA represent relatively large losses
in the hand, but week in the head.
This paragraph aims to choose the model of the PIFA antenna and its parameters.
The structure of antenna consists to choose the metal of the plate which induce
a short circuit between the radiating patch and the ground plane, as in the Fig. 2,
also the metal of (substrate, patch, the ground, the feed rod, the coaxial cable) [10]
(Table 1).
And to improve the performance of the PIFA antenna as well as to solve the
problem of having a narrow bandwidth there are several techniques:
• Increase the height of the substrate to increase the volume of the antenna.
• Reduce the size of the ground plane.
• Inserts slots in the round plane.
• Use of material of low relative permittivity as substrate.
• Adding a parallel plane between the patch and the ground plane (load capacitive).
• Inserting slots into the patch.
1042 S. Arsalane et al.
For the antenna presented in this paper we chose to insert slots in the patch to
have a multiband antenna.
And as the dimensions of the antenna had an impact on the directivity of the
antenna, so it is important to choose it carefully (Table 2).
After determination of all the factors that will be used to simulate the antenna and
also the dimensions of PIFA, so this paragraph will be concentrated on the conception
and simulation of the antenna in the chosen frequency range (2 GHz–6 GHz).
The calculations are made in far field (Fraunnhofer area) (Fig. 3).
Concerning the adaptation of the antenna, the following figure shows the
magnitude of the reflection coefficient in the frequency range (2 GHz–6 GHz) (Fig. 4).
The measurements are made in the frequency range (2 GHz−6 GHz) which
includes the second frequency range of the 5G technology.
The magnitude of the reflections coefficient is in decibel with a frequency up to
6 GHz. As we can notice that the magnitude of reflection coefficient S (1, 1) is less
than (−10 dB) for three frequencies (2.4 GHz, 3.4 GHz and 5.4 GHz) which means
that our antenna is a tri band antenna, and it is adapted for this three frequencies
and as a consequence we have the transmission of the maximum of electromagnetic
radiation.
For the radiation pattern that it obtained by the calculation of the electric field
in two plane E and H. that it represent the energy radiated by an antenna. It’s a
Design and Simulation of Three Bands Planar Inverted F Antenna Array … 1043
For the opening angle it fluctuate between 60 to 120 theta but it is intensive from
50 to 100, so the opening angle is 50 theta in the far field.
After the simulation of our antenna and obtaining the different results, we will
approach the periodization of this antenna in order to obtain an antenna array, which
will be functioning in the 5G frequency band and used subsequently in tests as an
equipment under tests.
Design and Simulation of Three Bands Planar Inverted F Antenna Array … 1045
PIFA or formed beam antennas are used for several purposes, an antenna array is
a collection of separate antennas and fed synchronously, the electromagnetic field
produced by an antenna array is the vector sum of the fields produced by each of the
elements.
In addition these antennas are generally not very cumbersome in comparison with
horn type aperture antennas or reflector antenna which would fulfill the same mission
[13].
Principle
Antenna array are the result of the judicious association of several antennas to
synthesize a radiant opening [14].
Antenna array are made up of N radiating sources (dipoles, patches, PIFA…)
distributed in space. The electrical quantities (signals) injected or collected on the
access of each of these source are weighted in amplitude and (or) in phase before to
be summoned among themselves. A beam splitter is used to perform coherent and
weighted summations of the signals collected by a receiving systems, and the splitter
will be injected into each source of antenna. Due to this weighting, the antenna arrays
can produce diagrams of radiations having the desired shape [15]. In Particular, it
will be possible to create several lobes simultaneously or a lobe in the direction of
the incident signal and a zero in the direction interference.
Array antennas can have several geometric configurations that can be grouped as
follows:
• Linear Arrays: alignment of sources on a straight line.
• Planar Arrays: the sources are arranged on a plane.
• Circular Arrays: the sources are arranged on a circle.
• Volume Arrays: the sources are distributed in a volume.
Several factors contribute to the formation of the antenna radiation pattern:
• The geometric configuration (linear, circular, planar, volume).
• The distance between the sources.
• The amplitude of excitation for each element.
• The polarization phase for each element.
For our case we choose to implement a linear array, it is the most frequently used
geometry in the design of antenna arrays because it is easy to implement, it is made
up of N sources equidistant with a lattice patched. These sources are supplied with
the same amplitude and with a progressive phase gradient for a point M located in
the far radiation zone, the total radiated field is given by the sum of each elementary
field [16].
Therefor this paragraph present the case of PIFA array in a linear geometric. The
linear array composed of four elements (PIFA antenna) distributed on an axis and
separated by a distance d (pitch of the array) (Fig. 7).
1046 S. Arsalane et al.
And for the 3D presentation it shows the concentration of the energy in a specific
direction, that represent to the width of the major lobe which is determinate by the
angular width of each boundary of the lobe whose intensity decreases by half, a
decrease of 3 decibel.
1048 S. Arsalane et al.
For the energy balance of the antenna which means the radiated power carried by
the electromagnetic wave for a specified frequency. The energy of each frequency it
presented in this table (Table 3):
The radiated electric energy depends on the frequency, and that what we observe
in the table, the energy balance increases with the frequency.
3 Conclusion
References
12. Luo Y, Zhang Z, Chen H, Huangfu J, Ran L (2009) High directivity antenna with small aperture,
November
13. Sayidmarie KH, Maewan NJ, Aabd R, See C (2019) Antennas for emerging 5G systems
14. Reckeweg M, Rohner C (2015) Antenna Basics
15. Kumar ST, Susila M (2017) Design of multi resonant PIFA for mobile telecommunication
networks
16. Rhee E (2020) Miniaturized PIFA for 5g communication networks, 15 February
17. Singh H, Sneha HL, Jha RM (2013) Mutual coupling in phased arrays, April
18. Abimbola RA (2017) Effects of spacing and number of elements on the fields radiated by linear
arrays of equally spaced elements, October
RFID System for Hospital Monitoring
and Medication Tracking Using Digital
Signature
Abstract The enhancement of patient safety and the improvement of medical care
services remain one the main concerns of the World Health Organization. Thus,
given the healthcare critical impact on patients, there is a growing need for high-risk
and expensive medicine management. For this purpose, we propose an innovative
system prototype for the reduction of medication error possibility from prescription,
validation to its preparation. We present a Radio Frequency IDentification monitoring
system for hospital and pharmacy using digital signature. Our designed system is
constituted of RFID readers, ESP8266 modules, and tags to ensure an automated
medication management and a good inventory control.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1051
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_96
1052 S. El Abkari et al.
Many studies on medication monitoring systems have evolved and achieved high
performances by utilizing different wireless technologies. Table 1 below present
different wireless identification technology related management systems.
constituted of a tag, a reader and a software application [15]. RFID tags include chips
for data storage and antennas while RFID readers (responders) read and can write
information of multiple tags simultaneously without line-of-sight communication
[16, 17]. RFID technology offers great advantages such as its reusability, ease of
data transmission, and security [18]. Figure below illustrates the RFID identification
process which consists of reading an RFID tag attached to a person or an object
without any physical contact (Fig. 1).
1054 S. El Abkari et al.
3.3 Methodology
Our RFID management system consists of four main parts (Fig. 3):
• Pharmacy system (Fig. 2. b) placed in the pharmacy,
• Nurse/doctor modules (Fig. 2. a) carried by the medical staff,
• Tracking modules placed at each room door,
• Monitoring system which collects and processes data collected using RFID and
Wi-Fi devices.
For the pharmacy system, we establish communication between RFID module and
Arduino Uno board with Serial Peripheral Interface (SPI) using SS, clock, MISO,
and MOSI. ESP8266 module and Arduino Uno board [20] communicate with TX
and RX pins. Thus, ESP8266 modules are utilized to send RFID IDs to pharmacy
and other system components to track medications.
RFID System for Hospital Monitoring and Medication Tracking … 1055
Medications (Fig. 4), patients, nurses, and doctors have their respective tags and
IDs (Fig. 5). Medical staff and medications are tracked and monitored by PN532-
NFC and ESP8266 modules. Each patient, doctor, and nurse have a unique digital
signature (unique identification code (ID)).
Traditionally, prescriptions are handwritten, nurses verbally inform the pharma-
cist about a patient medication. In addition, records of token medication are done
manually by the pharmacist (Table 3). It means that it is mainly based on human
resources which leads to many errors during the medication process. Our system
attempts to locate, manage, and monitor medications by improving the traditional
practices through an automated monitoring system. Tables 3 and 4 show traditional
and proposed medication monitoring processes.
The medication process is constituted of two main phases: Prescription phase and
medicine dispensing-tracking phase.
• Prescription phase
1056 S. El Abkari et al.
In this phase, the doctor prescribes medications after diagnosing a patient. The
pharma-cist then receives prescriptions signed with doctors’ IDs, verify IDs,
prepares medications, and sends a notification to the nurse in charge (Fig. 6.).
If there is an error in the prescription, the doctor receives a query to update it.
1058 S. El Abkari et al.
Fig. 7 Prescription
information received by the
nurse
Once the nurse receives medications prepared by the pharmacist, the monitoring
system starts tracking those medications. Tracking modules send then notifications
of medication locations with correspondent nurse ID to the monitoring system.
ESP8266 modules are also placed in different places in the hospital to permit
communication and data transfer within the network.
Our proposed system shows an improvement in patients’ safety in hospitals and
the medication process. Table 5 below illustrates features of our proposed system
compared to existing medication monitoring systems.
RFID System for Hospital Monitoring and Medication Tracking … 1059
4 Conclusion
References
1. Husar J, Iakovets A (2019) Proposal for monitoring the movement of biological samples in
hospitals. In: Automation and Management in Theory and Practice. In: 13th Anniversary Con-
ference of Universities, High Schools and Practice, Stará Lesná, SR
2. Balog M, Husár J, Iakovets A (2020) RFID System for Effective Managing of Medical Facilities.
New Approaches in Management of Smart Manufacturing Systems, pp 1–23. Springer, Cham
3. Mhamdi J, El Abkari S (2015) Contriving an RFID system for Alzheimer patients tracking.
In: 3rd International Workshop on RFID and Adaptive Wireless Sensor Networks, pp 23–28.
IEEE, Agadir
4. Bates DW, Cullen DJ, Laird N, Petersen LA, Small SD, Servi D, Vander Vliet M (1995)
Incidence of adverse drug events and potential adverse drug events: implications for prevention.
JAMA 274(1):29–34
5. Omotosho A, Ayegba P (2019) Medication adherence: a review and lessons for developing
countries. Int J Online Biomed Eng (iJOE) 15(11):104–123
6. Koutkias VG, Chouvarda I, Triantafyllidis A, Malousi A, Giaglis GD, Maglaveras N (2009)
A personalized framework for medication treatment management in chronic care. IEEE Trans
Inf Technol Biomed 14(2):464–472
1060 S. El Abkari et al.
7. Tejesh BSS, Neeraja SJAEJ (2018) Warehouse inventory management system using IoT and
open-source framework. Alexandria Eng J 57(4):3817–3823
8. Balog M, Matiskova D, Semanco P (2017) System of Workpiece Automation with RFID
Identifier Published Patent Application SK 50063–2016 Darina Matisková, A., Michal Ba-log,
Pavol Semančo Banská Bystrica: IPO SR 7
9. Sarzynski E, Decker B, Thul A, Weismantel D, Melaragni R, Cholakis E, Tewari M, Beckholt
K, Zaroukian M, Kennedy AC (2017) Beta testing a novel smartphone applica-tion to improve
medication adherence. Telemed e-Health 23:339–348
10. Morak J, Schwarz M, Hayn D, Schreier G (2012) Feasibility of mHealth and Near Field
Communication technology-based medication adherence monitoring. In: Annual International
Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pp 272–275.
IEEE
11. Yang J, Feng Z, Ma X, Zhang X (2018) Indoor positioning method based on wireless signal.
Int J Online Biomed Eng (iJOE) 14(10):53–67
12. Schreier G, Schwarz M, Modre-Osprian R, Kastner P, Scherr D, Fruhwald F (2013) Design and
evaluation of a multimodal mHealth based medication management system for patient self-
administration. In: 35th Annual International Conference of the IEEE Engineering in Medicine
and Biology Society (EMBC), pp 7270–7273
13. Mirea A, Albu A (2018) Acquisition of physical data in an automated system for monitoring
medication stocks. In: 12th International Symposium on Applied Computational Intelligence
and Informatics (SACI), pp 000179–000182
14. Khan SF (2017) Health care monitoring system in Internet of Things (IoT) by using RFID. In:
International Conference on Industrial Technology and Management (ICITM), pp 198–204
15. McCarthy JF, Nguyen DH, Rashid AM, Soroczak S (2002) Proactive displays and the
experience UbiComp project. ACM SIGGROUP Bull 23(3):38–41
16. RFID Journal, 2002–2007, “referred 6.8.2007” https://www.rfidjournal.com. Accessed 25 July
2020
17. El Abkari S, Jilbab A, El Mhamdi J (2018) Real time positioning over WSN and RFID network
integration. In: 4th International Conference on Advanced Technologies for Signal and Image
Processing (ATSIP), pp 1–5. IEEE, Tunisia
18. Cheng-Ju L, Li L, Shi-Zong C, Chi Chen W, Chun-Huang H, Xin-Mei C (2014) Mobile
healthcare service system using rfid. In: International Conference on Networking, Sensing and
Control, vol 2, pp 1014–1019
19. Castro L, Lefebvre E, Lefebvre LA (2013) Adding intelligence to mobile asset management
in hospitals: the true value of RFID. J Med Syst 37:1–17
20. El Abkari S, Jilbab A, El Mhamdi J (2020) Wireless Indoor localization using fingerprinting
and Trilateration. J Adv Res Dyn Control Syst 12(7):2597–2602
21. Devos P, Jou AM, De Waele G, Petrovic M (2015) Design for personalized mobile health
applications for enhanced older people participation. Eur Geriatr Med 6:593–597
22. Abbey B, Alipour A, Gilmour L, Camp C, Hofer C, Lederer R, Rasmussen G, Liu L, Nikolaidis
I, Stroulia E (2012) A remotely programmable smart pillbox for enhancing med-ication adher-
ence. In: Computer-Based Medical Systems (CBMS). In: 25th International Sym-posium, pp
1–4
23. Neubeck L, Coorey G, Peiris D, Mulley J, Heeley E, Hersch F, Redfern J (2016) Devel-opment
of an integrated e-health tool for people with, or at high risk of, cardiovascular dis-ease: the
consumer navigation of electronic cardiovascular tools (CONNECT) web application. Int J
Med Informatics 96:4–37
24. Ferreira F, Almeida N, Rosa AF, Oliveira A, Teixeira A, Pereira JC (2013) Multimodal and
adaptable medication assistant for the elderly: a prototype for interaction and usability in
smartphones. In: Information Systems and Technologies (CISTI), pp 1–6
Building a Domain Ontology
for the Construction Industry: Towards
Knowledge Sharing
Abstract As the construction industry has crucial interplays with many related
fields, all in a constant state of flux, the use of a domain ontology to capture various
relevant entries from related subdomains has become of paramount importance. The
purpose of this article is to conceptualize the utility of ontologies in the construction
industry as a solution for inter-sectorial transactions. Thereafter, upon presenting the
process of ontology development, a real-life company was taken as a case study to
implement the developed ontology. The purpose behind this study is to fully automate
the generation of a domain ontology from Excel to OWL files. The methodology and
process followed are thoroughly explained in the paper. The process is meant to
be generic in such a way it could be reused in other projects for other industries,
provided that the data is hierarchical and presented under Excel.
1 Introduction
The construction industry is constantly evolving, and new technologies are being
implemented and used on a regular basis. One of those trendy technologies is ontolo-
gies. In fact, because the construction industry generally implicates stakeholders
from different backgrounds and disciplines, having a “common representation of
concepts” is crucial and has proven to be effective and time-saving.
The use of ontologies in the construction industry is gaining popularity. Indeed,
their ability to describe relationships and their high interconnectedness make them
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1061
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_97
1062 H. El Asri et al.
the bases for modeling high-quality, linked and coherent data [1]. Research in this
area has been going on since the 2000s; however, it was not as advanced as it is today.
In fact, before 2000, research was focused on Artificial Intelligence (AI) and software
requirements. After the 2000s, AI was still in vogue, but several technologies emerged
like ontologies that became more popular with the rise of knowledge management.
Further, with the emergence of BIM (Building Information Modeling) technologies
[2], the use of ontologies has become even more important. However, the discrepancy
in file and item labelling is a seeming problem that the solution proposed in this article
tries to solve.
This paper is structured as follows. Section 2 discusses key concepts in ontolo-
gies. Section 3 discusses some relevant previous works in the field of construc-
tion. Section 4 presents the case study for which a general ontology was elaborated.
Section 5 discusses the domain ontology implementation with a focus on the mapping
rules and the implementation process. Last but not least, Sect. 6 presents the future
works.
2.1 Definition
The definition of ontologies evolved tremendously over the past decades. In 1995,
Gruber defined ontology as an “explicit specification of a conceptualization”. A
few years later, in 1998, [1] added that ontologies can be used in many applica-
tions to “capture relationships, boost knowledge management, and make knowl-
edge assets intelligently accessible to people in organizations”. In 2001, the defi-
nition evolved and started encompassing conceptual models and implementations.
[3] defined ontologies as the implementation of a conceptual model that satisfies
the engineering trade-offs of a running application. In the past few years, ontolo-
gies have started being linked to knowledge management. [4] said that “Ontologies
play an important role in knowledge management and representation. In practice,
ontology means a hierarchy of concepts with a set of properties and relations that
represent a domain. Properties of each class describe the features and attributes of the
class and their restrictions.” This is, however, not always true as it is not a hierarchy.
What is interesting about ontologies is the inference process.
In a nutshell, ontologies can be seen as a “unifying structure” that enables
researchers to represent information in such a way that it provides common under-
standing and representation [5]. Furthermore, ontologies can represent any informa-
tion. They are often associated with taxonomic hierarchies of classes, although they
are not limited to this only.
Building a Domain Ontology for the Construction Industry ... 1063
Different types of ontologies exist and each is used to serve a specific purpose.
The following are the well-known types. General ontologies are known to contain
and provide general knowledge rather than domain-specific knowledge. Domain
ontologies that are focused and applicable to a specific application domain, which
helps in defining how a group of users conceptualize and visualize some specific
phenomenon. Core reference ontologies that are usually linked to a specific domain;
yet, integrates different viewpoints related to groups of users. It is considered to be
the result of numerous domain ontologies. Information ontologies focus on concepts,
instances, and their relationships with the aim of proposing a general overview.
Last but not least, linguistic or terminological ontologies that can be glossaries,
dictionaries, controlled vocabularies, taxonomies, folksonomies, thesaurus, or lexical
databases. According to [6], “the roles of linguistic ontologies are twofold: The first
one is to present and define the vocabulary used. Second, linguistic ontology is the
result of a terminology agreement between a users’ community. This agreement
defines which term is used to represent a concept in order to avoid ambiguity.”
It is worth noting that an ontology and a thesaurus are completely different in
several aspects. On the one hand, ontologies are multidimensional and flexible.
They provide “controlled vocabulary” to describe a specific domain in a hierar-
chical manner. Furthermore, ontologies do not only provide definitions, but also
relationships, axioms, and restrictions. On the other hand, a thesaurus is rigid and
single-dimensional. It does not provide definitions, but rather relationships. It should
be noted that there exist three types of relationships only: BT (Broader Term), RT
(Related Term), and NT (Narrower Term).
2.3 Tools
Several languages, mechanisms, and tools are available for developing, constructing,
and implementing ontologies. The main languages used are: XML, RDF, RDFS, and
OWL. Many tools based on the aforementioned languages have been developed and
widely used like Protégé, WebOnto, WebODE, Swoop, and OWLGrEd, to mention
but a few.
1064 H. El Asri et al.
According to [8], there are 5 main phases for building and implementing an
ontology: methodology, pre-development, development, implementation, and post-
development. The first phase, the methodology, is all about the extraction of the
domain-specific terminology from the different documents available and from the
web. In the pre-development stage, a bottom-up conceptualization is used to construct
the core ontology, which will define the semantics of the referential ontology. In this
step, a conceptual graph is also formalized. In the development phase, all requirement
specifications are analyzed and the taxonomic structure is elaborated. Representing
the ontology using a specific language, OWL for example, is carried out and the
conceptual model is officially elaborated. Then, the implementation process starts
and is focused on two main aspects: the operations and the exploitation. In the oper-
ationalization process, making the core ontology context-specific is key. Then, in the
exploitation process, the focus shifts towards integrating the referential ontology into
a knowledge based-system that includes a fact base and a rule base. Maintenance
happens in the post-development phase. Last but not least, determining the need for
change after implementation happens in the integration process.
For this project, we are interested in developing and implementing a semi-
automatic process that would enable the generation of domain-specific ontologies
for the construction industry.
3 Previous Works
In the past two decades, researchers got interested in ontologies and have started to
investigate their uses and how they can help the construction industry move forward.
For instance, [9] proposed a systematic domain ontology for requirements change
management process in the global software development environment. A hybrid
method combining methontology, which is a well-known approach, and the 101
method was used for its development, the Web Ontology Language (OWL) for its
representation, and Protégé for its implementation. According to the researchers,
there are several benefits for building such ontology; for example, it can be used
to mitigate miscommunication and misunderstanding issues that may arise within
organizations.
Besides, [10] presented a method for developing an ontology of web-pages
mapped to a domain knowledge. The researchers focused on solutions of semi-
automatic ontology construction using web usage data; an experiment of Microsoft
Web data was implemented and evaluated. The goal behind this study, that was
successfully achieved, was to show the efficiency of page recommendations based
on a domain ontology. Similarly, [11] have also conducted a study on ontologies as
a crucial element in the semantic web. An ontology for the health tourism industry
in Thailand was implemented. The final results have shown that the ontology was
efficient and further works are getting implemented.
From another perspective, [12] presented a comprehensive analysis of the align-
ments between one domain ontology from the OAEI Conference track and three well
known top-level ontologies, namely: DOLCE, GFO and SUMO. The researchers
mentioned that the reuse of the knowledge available in top-level ontologies along
with that of domain ontologies reduces the modelling time and complexity, hence the
heterogeneity issue of knowledge representation. The problem of matching domain
and top-level ontologies has been addressed and a viable solution was proposed.
In addition, [13] investigated the conversion from EXPRESS, a schema that
defines the standard IFC (Industry Foundation Classes), into OWL for the construc-
tion industry. Although this has previously been investigated by researchers, no
results turned into a referential yet. [13] analyzed the available research and came
up with recommendations in order to get a usable ontology. The result of this study
is a list of criteria and requirements to be followed in order to generate a referential
ontology.
Furthermore, [14] investigated an approach for building domain ontologies for
guiding requirements’ elicitation and extracting semantic graphs from textual tech-
nical standards to generate compatible baseline domain ontologies. To achieve that,
three main steps were followed: term extraction, term clustering, and taxonomy
building. The researchers identified three core properties of domain ontologies
that they called “explicit relational expression, qualified relation identification, and
explicit temporal and spatial expressions.” The results of the paper suggest that the
methodology/approach presented in the paper could help reduce the hassle that comes
with building a domain ontology from scratch.
The number of research papers available that discuss ontologies in the construction
industry are not that numerous. Those that investigated automation processes are
limited. What we are proposing in this article is to provide an automated process that
could be reused in other domains where data is hierarchical in Excel files.
1066 H. El Asri et al.
4 Case Study
A Rabat-based construction company was taken as a case study. This company uses
many documents (mainly in excel, word, and pdf formats) that are shared among
several collaborators. As an example, Fig. 1 presents a snapshot of one of the
numerous documents used by the company, which is in Excel format: the “bor-
dereau”. It is one of the main documents used in all the works that the company
does.
The column “designation des ouvrages” (2nd column) represents the different
classes and subclasses while the column “N° des prix” (1st column) provides the
identification of each class/subclass. For instance, the class “Gros oeuvre – Charp-
ente Métallique” is a main class that contains 2 direct sub-classes, namely: “terrasse-
ments et remblais” (identifier 100.1) and “bétons et acier en infrastructure” (identifier
100.2). Each of these sub-classes has several sub-classes of its own.
This document is the basis of an important process in the company and the problem
that can occur is that collaborators sometimes use different names for the same
concept. Therefore, a domain ontology is of great help. Our objective was then to
develop a domain ontology based on several documents provided.
The main objective behind developing this domain ontology for the construction
industry is to provide a common and unified terminology to enable the experts who
work on the same project(s) to use the same terminology.
UNITE DES
N° DES PRIX DESIGNATION DES OUVRAGES EN TOUTES LETTRES
MESURES
100 GROS-ŒUVRE - CHARPENTE MÉTALLIQUE
100.1 TERRASSEMENTS ET REMBLAIS
100.1.1 TERRASSEMENTS
100.1.1.1 FOUILLES EN PLEIN MASSE TOUT TERRAINS m³
100.1.1.2 FOUILLES EN PUITS ET EN RIGOLE TOUT TERRAINS m³
100.1.2 REMBLAIS
100.1.2.1 REMBLAI D'APPORT EN TOUT VENANT COMPACTE m³
100.1.2.2 REMBLAIS EN MATERIAUX PROVENANT DES FOUILLES m³
100.1.2.3 ÉVACUATION DES DÉBLAIS m³
SOUS-TOTAL HT DE TERRASSEMENTS ET REMBLAIS
100.2 BÉTONS ET ACIERS EN INFRASTRUCTURE
100.2.1 BÉTON DE PROPRETÉ m³
100.2.2 GROS BÉTON m³
100.2.3 BÉTON ARMÉ EN INFRASTRUCTURE
100.2.3.1 BÉTON POUR SEMELLES ISOLÉES, FILANTES ET RADIERS m³
The ontology development was divided into two main phases: the specification phase,
in which the domain knowledge is acquired, and the conceptualization phase, in
which the knowledge is structured and organized. Figure 2 presents the general
methodology followed to build this ontology.
The process started with the gathering of the user requirements from the company.
A thorough analysis was conducted and a pre-building phase started. To make sure
the pre-building phase was heading in the right direction, several interviews and
document checks were carried out as part of the user requirements gathering. After
making sure that all the specifications were taken into account, the ontology develop-
ment and implementation was carried out. Last but not least, an ontology evaluation
was conducted.
Because the majority of the company’s data is available in Excel files, our goal
was to automate the domain ontology implementation process to make it easier, time-
saving, and reliable to have access to the needed data/information. The first step was
to find the mapping rules.
The use of ontologies in the construction and building industry is growing. Because
of the multi-participants from different areas of expertise and disciplines that work in
the construction industry, and the documents that are usually written and checked by
different ex-perts coming from various backgrounds, having a uni-fied representation
of the general terms used can be time-saving and of great help. This paper presented
a domain ontology developed for a specific construction company, following their
requirements and using their data based on a semi-automatic process. At the moment,
we are working on the implementation of the entire process using Java, Apache POI
for reading the Excel files, and Jena to generate the OWL file. Furthermore, and
since its apparition, Building Information Modelling (BIM) has been considered as
a shifting point in sharing and representing knowledge in the construction industry.
Although not widely used yet, companies and researchers have realized the impor-
tance of all the benefits that come with using it. Integrating ontologies within the BIM
environment can only be beneficial and will add to the advantages that BIM already
offers. It will allow the creation of a “common jargon”, which will help in standard-
izing the meaning concepts and terminologies, encourage knowledge sharing, and
promote the reuse of existing knowledge. For this project, researchers are currently
working on the integration of the domain ontology into BIM. Last but not least,
Building a Domain Ontology for the Construction Industry ... 1071
ontology development and implementation is a continuous and long process, and the
next step will be to develop a full-fetched domain ontology, to be integrated in BIM.
References
Abstract This article presents a hybrid genetic algorithm (GA) with two-
optimization algorithm, to find a solution to the problem routing in Ad Hoc network
big map, the objective is to find the shortest possible distance between a trans-
mitter head (cluster, node, intermediate or advanced) and the base station. From
studies and remarks, it has been proved that in the topologies of the large Ad Hoc
network, the node responsible for transmitting to the base station sometimes used
up its total energy and became inactive. So, there are many techniques to solve the
problem routing in Ad Hoc networks, like GA (Genetic Algorithm), SA (Simulated
Annealing), and BA (Bat Algorithm). In this article, we are focus on the (GA) that
are present in the basic approaches of our work. We propose an analysis of these
metaheuristics by identifying the fundamental ideas and propose ways to guide the
choice of a metaheuristic in practice, particularly in Ad Hoc big map network, to
find the best path between two nodes in communication. Thereafter, we try to more
optimizer our algorithm by using a 2-optimizer algorithm, and we compare between
two method GA and GA-2opt, the result show much better quality of our solution to
adjust the problem of routing in big map Ad Hoc network.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1073
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_98
1074 H. Khankhour et al.
Then, this new population undergoes a selection according to the fintess of its indi-
vidual, that means that only the individuals best adapted to the environment survive
[1, 2].
In Ad Hoc network we have considered chromosomes as routes (Strings) which
are variable, and their genes as nodes (parameters) such that the distance between
nodes is different to zero, first we randomly generalize a population, which contains
several chromosomes (path), we used the single point crossing by chance (0.8), to
better improve the quality of our solution we added the mutation operator (0.05)
from the last population and we evaluated each chromosome (path) by the function
of fitness (1), the best solution must be chosen from a source node and a destination
node in Ad Hoc Network.
1
f i = li−1 (3)
j=1 C gi( j),gi( j+1)
Fig. 2 Example of
permutation in algorithm
with two-optimization
In 1958, the two optimizations (2-opt) algorithm has been proposed by Cross to solve
the traveling salesman problem. The ultimate idea is to take a road that is crossing
and rearrange it [10] (Fig. 2).
2-optimization is an iterative algorithm: at each step, we remove two edges from
the current solution and reconnect the two turns thus formed. This method makes
it possible to improve the solutions by avoiding the path when tour inequality is
respected between a source node and a destination node (see figure above). On the
Optimization of the Ad Hoc Network by Using Hybridization … 1077
diagram on the right, the route (a ->b->e->d->c->f->g) is modified to the route (a-
>b->c->d->e->f->g) by reversing the order of visit to tour e and c. More generally,
reversing the running order of two tours need to reverse the running order of all the
tours between them [11].
In this paper, our new GA-2opt approach aims to solve the routing problem in Ad
Hoc network, between the source node and the destination node in a short time too
[12], we have proposed a new algorithm design as follows:
To see the effect of the 2-optimizer algorithm on the Ad Hoc network, we have built
a program that uses the uniform (GA) as a heuristic, and in the crossing step we
check both children by the “2- optimizer”, if there is an optimization, replace the
child (route) with the new solution.
Table 1 shows the great effect of this optimize.
From the results in Fig. 3, we notice that the distance of the genetic algorithm 2-
opt is much less than the (GA) mostly for the big city like Bier127. it means that
Optimization of the Ad Hoc Network by Using Hybridization … 1079
in a big map network using 2-opt optimization we can find the shortest path in the
shortest time too, from the source node to the destination node.
After testing our algorithm with the BENCHMARK instances, we applied our
algorithm to the ad hoc network using the topology Fig. 4, used in this article [9]
after several execution of our program, from the above results we can say that the
delay for our new design algorithm AG2OPT is good, we find the best path, the cost
is 142, in short time too 0,020 s better than result in this article [9].
8 Conclusion
This article proposes a new conception of the hybridization between the metaheuristic
(GA) algorithm and two optimization algorithms, and we choose to further optimize
our algorithm using the advantages of the 2 opt optimizer, for this purpose the goal is
to find the shortest path from a source to a destination in quick time too, mainly for
a large network of cardswhose goal that the nodes do not lose its energy, choosing
the best path, to avoid shocks, collision and of course not to saturate the bandwidth
in Ad Hoc network in a big map.
1080 H. Khankhour et al.
References
1 Introduction
The role of an antenna is to convert the electrical energy of a signal into electromag-
netic energy, or conversely to convert electromagnetic energy into electrical energy.
A transmitting antenna is a device that ensures the transmission of energy between
a transmitter and the free space where this energy will propagate. Conversely, a
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1081
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_99
1082 S.-E. Didi et al.
receiving antenna is a device that transmits the energy of a wave propagating in space
to a receiving device. The patch antenna is a flat antenna whose radiating element is
generally a square conductive surface, separated from a conductive reflector plane by
a dielectric substrate. Its realization resembles a double-sided printed circuit board,
substrate, and is therefore suitable for industrial production. The use of patch antennas
is almost generalized in all mobile communication systems. These antennas are light,
compact, and inexpensive [1].
The 5th generation (5G) wireless technology network is scheduled to be deployed
by 2020. It is the most ambitious wireless communication network project ever
carried out and will have to meet the expectations of a mass market. The latest spec-
ification trends focus on network densification, increased data throughput (10 Gb/s,
10 times higher than 4G), increased network coverage, latency between 1 ms and
5 ms and reduced power consumption [2].
Frequency bands: It is planned to free up frequencies for 5G that are located prac-
tically all along the radio spectrum. The National Frequency Agency (ANFR) indi-
cates that they are divided into two major classes. The low frequencies offer extensive
coverage and excellent transmission in the indoor environment. For high frequen-
cies, they are strong, but the transmission in residential environments is restricted.
Therefore, the chosen technology strategy is to combine the new radio spectrum as
well as frequencies currently available for 2G, 3G, and 4G mobile data transmission.
Currently, Two new funding tranches will apply to 5G: the 3,5 GHz frequency band is
between 3,4 GHz and 3,8 GHz and the 26 GHz frequency band is in the [24,25 GHz;
27,5 GHz] frequency range [3, 4].
In this article, we present a patch antenna with two slots for 5G. This proposed
antenna must resonate at 27.5 GHz and has a size of 4.678 mm × 4.55 mm ×
0.15 mm structure. In the first phase, we introduce the required introduction, in the
second phase; we explain the design methodology of this antenna offer, in the third
phase, we carry out the simulation of this created antenna through an HFSS (High-
Frequency Structure Simulator) software tool. The details of these simulations will be
discussed, and we thus compare the different results acquired from this new antenna
with those already available. In the fourth phase, we finish this work by concluding
with a conclusion.
In this paragraph, we present the most commonly used method for designing
microstrip patch antennas, including the procedure for determining their parame-
ters (width and length), and also the dimensions of the ground plane, and those of the
feed lines, and the type of power supply for this design, as well as how to glue the
patch antenna to the substrate of your choice. For simplified analyses and antenna
performance predictions, we propose a rectangular patch antenna with two slots oper-
ating at 27.5 GHz as part of the fifth generation 5G application, as illustrated in Fig. 1.
Design of a Microstrip Antenna Two-Slot for Fifth Generation Applications … 1083
The proposed antenna has been printed on a substrate of permittivity εr = 4.3. Poly-
imides are commonly used for coating metal substrates, and use a microstrip line for
their feed. Quarter Wave Converter is a microstrip cable employed to ensure proper
impedance matching. To determine the width and length of this quarter-wavelength
(WT, LT) and those of the micro-strip feed cable (Wf, Lf), and those of this offered
antenna (Wp, Lp), we apply the formulas in [5, 6].
• by applying Eq. (1), it is possible to find the value of the width Wp:
c 2 1
W p = 0.5 ( )2 (1)
f r εr + 1
• by applying Eq. (2), it is possible to find the value of the width Lp:
c
L p = L e f f − 2L p = √ − 2L p (2)
2 f r εe f f
where: the symbols Wp and Lp represent the width and length of the rectangular
patch antenna respectively, fr represents the working frequency, c shows that the
celerity of light is approximately equal to: 300000000 m/s, Lp is shown the stretch
path, Hs is the height of the material substrate, εeff indicates the set value of the
dielectric constant of the substrate material, WE is the latitude of the earth plane,
LE is the longitude of the earth plane (Table 1).
The return loss S11 of an antenna makes it possible to quantify the quantity of
the reflected signal in relation to the incident signal. This return loss is equal to −
10 of the fund value, indicating that about 10% of the received power is returned,
Design of a Microstrip Antenna Two-Slot for Fifth Generation Applications … 1085
that is to say that 90% as much power is absorbed through the antenna, something
that is believed to be very good for wireless mobile communication. the reflection
coefficient at the antenna level, if there is a perfect match there is no reflection, that
is to say that in this case the reflection coefficient will be equal to zero, if there is no
match the reflection coefficient is different from zero, we can define the qualification
at the match level of an antenna either by attributing its impedance characteristic,
generally equal to 50 ohms, or by attributing to its level of reflection coefficient. This
antenna functions at 26.91 GHz with a return loss of −25.4052 dB, a bandwidth of
0.96 GHz (27.35 GHz–26.39 GHz) as shown in Fig. 2 below.
VSWR occurs due to the reflection of a large amount of the generated power. The
value of this VSWR on the efficiency bandwidth must be between 1 and 2. In order
for the transmission of the antenna power to be maximized, this ratio must tend
towards 1. Figure 3 illustrates VSWR as a function of frequency. The VSWR value
is 1.095 at the operating frequency of this antenna.
3.3 Gain
The gain is equal to the ratio between the power density radiated by the antenna in
a specific direction and the power density emitted by a reference antenna. The gain
of this study for this new antenna is 6.6 decibel, as shown in Fig. 4 below. The gain
of this antenna obtained is quite suitable and its use is the most appropriate for 5G.
1086 S.-E. Didi et al.
Table 2 Comparative analysis of the results of this antenna in relation to those of the existent
results
3.4 Directivity
Fig. 5 The
three-dimensional directivity
pattern of the offered
microstrip antenna
1088 S.-E. Didi et al.
4 Conclusion
In this article, I have shown the definition and characteristics of the antenna as well
as some of the advantages of the fifth generation, these advantages being the date of
its publication, the increase of the data transmission rate, the increase of the coverage
area, the decrease of the power consumption, also the steps of the transition from the
first to the fifth generation, He also mentioned the presentation of the fifth generation
frequency bands. I specified the steps in the design of a patch antenna as well as the
study of the simulations on the HFSS software and I compared the results obtained
with those of some other antennas designed during work carried out in previous
years.
In this article, rectangular microstrip antenna is presented with two slots S1 and
S2 that improve antenna performance. The improved bandwidth allows for increased
user occupancy, indispensable for 5G portable communications. The global dimen-
sions of the offered antenna are 4.55 mm × 4.678 mm × 0.115 mm. The antenna
described in this document operates over 26,91 GHz and a reflection coefficient of −
25.4052 dB, a bandwidth of 0.96 GHz, a gain of 6.6 dB, and a directivity of 6.9 dB.
Through this study, we have shown that the two-slot antenna obtains the best
results in the relevant frequency band (24,25 GHz–27,5 GHz) on which it can work
well with the fifth generation portable communication system. The proposed antenna
would therefore be a good candidate for wireless applications at 27.5 GHz.
For future work, we will use this antenna in the design of patch antenna arrays to
improve the performance of patch antennas that meet the requirements of the fifth
generation 5G.
References
1. Odile P (2009) Antennas Theory, design and application. French center for the exploitation of
coupier law, Paris. https://www.dunod.com/
2. Faisal MMA, Nabil M, Kamruzzaman Md (2018) Design and simulation of a single element
high gain microstrip patch antenna for 5G wireless communication. In: Faisal MM (ed) Inter-
national Conference on Innovations in Science. Engineering and Technology (ICISET). IEEE,
Chittagong, pp 290–293
3. Iftikhar A, Sun H, Qasim A, Abdul S (2020) Design of umbrella shape single element patch
antenna with high gain and high efficiency for 5G wireless communication in 28 GHz. In:
17th International Bhurban Conference on Applied Sciences and Technology (IBCAST), pp
710–713. IEEE, Islamabad, Pakistan
4. Muhammad Z, Sultan S, Yasar A, Yuriy V, Peter E (2020) Ultra-wideband circular antenna
for future 5G. In: Conference of Russian Young Researchers in Electrical and Electronic
Engineering (EIConRus), pp 2280–2283. IEEE, St. Petersburg and Moscow, Russia
5. Ghanendra K, Chakresh K (2020) Design and analysis of u-slot mircostrip patch antenna for
mobile communication at 60 GHZ. Bull Eng
6. Dennis NA-C, Jorge LA-C (2018) High Gain 4 × 4 rectangular patch antenna array at
28 GHz for future 5G applications. In: XXV International Conference on Electronics, Electrical
Engineering and Computing (INTERCON), pp 1–4. IEEE. Lima, Peru
Design of a Microstrip Antenna Two-Slot for Fifth Generation Applications … 1089
1 Introduction
Nowadays, wireless technology requires antenna with wider bandwidth and smaller
dimensions. Therefore characteristics such as light weight, low profile, low cost,
easy to be manufactured and to be integrated with RF devices makes the Microstrip
antenna very attractive research area [1, 2]. Besides self-similarity and space fillings,
frequency independent, compact size make fractal antenna an excellent candidate to
minimize the size of the antenna or to improve its characteristics [2]. Furthermore
Fractal geometry is also combined with electromagnetic theory in order of investi-
gating a new class of radiation, propagation, and scattering problems. Recent studies
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1091
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_100
1092 F. Ez-Zaki et al.
on the microstrip fractal antenna have been recently reported in [3–11]. Different
fractal geometries, such as Hilbert, or Sierpinski curves, have been used to bring
the benefits of miniaturization [3], wide bandwidth [4–6], and multiband operation
[5–10]. So far the bandwidth of the microstrip fractal antenna is reported to be
narrow. Therefore multiple techniques are used to resolve this task such as using
slots, notches, defected ground structure (DGS) [12–14].
In this paper, we propose a novel truncated antenna based on the Sierpinski carpet
shape of the second order, and a slotted ground. The proposed antenna was simulated
using CST Microwave Studio. Impact of fractal iterations and defected ground plane
was analysed in details.
Patch length:
L = L e f f + 2L (5)
For the used frequency of 5.8 GHz the patch width and length are:
W = 15 mm and L = 12 mm, the substrate is of 20 × 20 mm2 , the dimensions of
the full ground plane are 20 × 20 mm2 . The thickness of substrate (h) is of 1.6 mm.
Antenna 2: four semi-circle corners are cut to provide as a perturbation as shown
in Fig. 2.
Antenna 3: To get 1st iteration of the antenna, the base patch geometry is divided
into 9 congruent sub-rectangles in a 3-by-3 grid, and remove the central rectangle as
shown in Fig. 3.
Antenna 4: the same procedure as step 3 each rectangle patch is divided of 1st
iteration geometry into 9 congruent sub-rectangles, and remove the central one as
illustrated in Fig. 4.
Antenna 5: To get the finale geometry a U-tree slots in the ground plane as shown
in Fig. 5.
Simulated S11 and VSWR versus frequency plots of all the antenna construction
stages are depicted in Figs. 6, and 7 respectively. The S11 for the eventual square
patch are less than −10 dB at 0.82, 5.543, 9.10, and 15.08 GHz frequencies whereas
the corner truncated patch antenna resonates at 0.84, 5.8, 9.39 GHz. The Sierpinski
corner truncated patch first iteration works at 0.87, 5.36, 12.35 GHz similarly to the
second iteration as it can be seen in Fig. 6 for antenna 3 and 4. So far the U-tree
slots Sierpinski corner truncated antenna shows more improved performances and
presents 7 resonant frequencies at 2.15, 4.37, 7.61, 11.3, 13.36, 15.743 GHz (Table 1).
So further the effect of the feed line dimensions is studied. As can be noted clearly
from the Fig. 8 and Fig. 9 the dimensions of the feed line have relatively smaller
effects on the antenna’s response.
As for the selection of the U-tree dimensions it is greatly dependent on the required
operating bands and they mainly affect the directivity of the antenna’s radiation
patterns. However, the U-tree dimensions are choose in order to get more improved
Performance Improvement of Corner-Truncated Sierpinski Carpet Fractal Antenna … 1095
antenna features and to attain the desired frequency band of vehicular communica-
tions at 5.9 GHz. In sum simulated results for S11 (dB) and VSWR of proposed
antenna are analysed in next section.
In this section, the simulation results of proposed antenna are analyzed in details.
Looking at Fig. 10, it is apparent that as we approach to antenna 5, the antenna reveals
more multiband and wideband behavior due to the increase in paths and lengths of
the antenna geometry by using the DGS technique. As it can be easily noticed from
the graph of the VSWR results the proposed antenna presents a good impedance
matching described by a VSWR closer to 1 within the operating bands.
1098 F. Ez-Zaki et al.
The simulated radiation pattern of the proposed antenna at resonant frequencies are
illustrated in Fig. 12. The designed antenna shows a nearly directional characteristics
in Phi = 90 with a very good directivity of a peak value of 6.87 dBi at almost operating
frequencies. As can clearly observed that the radiation patterns have multiple lobes
at higher frequencies than lower frequencies which leads that the radiation pattern is
distorted at high frequencies (Fig. 11).
Performance Improvement of Corner-Truncated Sierpinski Carpet Fractal Antenna … 1099
5 Conclusion
has been analyzed. The key features of this antenna is compact size, simple in
construction using Sierpinski fractal geometry. The investigated antenna works at
multiple frequencies (1.84, 4.09, 5.9, 7.18, 10.88, 13.17, 15.15, and 15.60 GHz)
and get improved performance for multi band (C, S, X and Ku) applications. Inves-
tigated antenna can be used for multiple applications and bands such LTE, WIFI,
and DSRC for vehicular communication, in addition to satellite, radar, and space
communications.
References
1. Werner DH, Ganguly S (2003) An overview of fractal antenna engineering research. IEEE
Antennas Propag Mag 45(1):38–57
2. Anguera J, Andújar A, Jayasinghe J et al (2020) Fractal antennas: an historical perspective.
Fractal Fractional 4(1):3
3. Singh K, Grewal V, Saxena R (2009) Fractal antennas: a novel miniaturization technique for
wireless communications. Int J Recent Trends Eng 2(5):172
4. Bhatt S, Mankodi P, Desai A et al (2017) Analysis of ultra wideband fractal antenna designs
and their applications for wireless communication: a survey. In: 2017 International conference
on inventive systems and control (ICISC). IEEE, pp 1–6
5. Soni BK, Singhai R (2018) Design and analysis of minkowskized hybrid fractal like antenna
for multiband operation. Prog Electromagnet Res 80:117–126
6. Ali JK, Abdulkareem S, Hammoodi A et al (2016) Cantor fractal-based printed slot antenna
for dual-band wireless applications. Int J Microwave Wirel Technol 8(2):263–270
7. Bhutani P, Sagar S, Kumar A (2019) Performance analysis of sierpinski carpet fractal antenna
for wireless communication. In: Applications of computing, automation and wireless systems
in electrical engineering. Springer, Singapore, pp 749–758
8. Dhankhar A, Kaur J (2015) Design of sierpinski carpet fractal antenna for multiband
applications
9. Radonic VK, Palmer GS, Crnojevic-Bengin V (2012) Flexible Sierpinski carpet fractal antenna
on a Hilbert slot patterned ground. Int J Antennas Propag 11:7 p, article ID 980916
10. Sivia JS, Kaur G, Sarao AK (2017) A modified sierpinski carpet fractal antenna for multiband
applications. Wireless Pers Commun 95(4):4269–4279
11. Wong S, Ooi BL (2001) Analysis and bandwidth enhancement of Sierpinski carpet antenna.
Microwave Opt Technol Lett 31(1):13–18
12. Gupta A, Joshi HD, Khanna R (2017) An X-shaped fractal antenna with DGS for multiband
applications. Int J Microwave Wirel Technol 9(5):1075
13. Arya AK, Kartikeyan MV, Patnaik A (2010) Defected ground structure in the perspective of
microstrip antennas: a review. Frequenz 64(5–6):79–84
14. Garg C, Kaur M (2014) A review of defected ground structure (DGS) in microwave design.
Int J Innovative Res Electr Electron Instrum Control Eng 2(3):1285–1290
15. Balanis CA (2005) Antenna Theory: Analysis and Design, 3rd edn. Wiley-Interscience,
Hoboken
A Novel System Based V2V
Communications to Prevent Road
Accidents in Morocco
Abstract Due to the deaths and injuries caused by road accidents every year, road
safety became an important field of research for many institutes. The development
of smart solutions can remarkably enhance road safety. Vehicular networks are one
of those solutions. It can bring advanced services, namely collision detection, traffic
management, communication between vehicles, and so on. This paper introduces an
implementation of a complete vehicle communication system. The prototypes are
developed using Raspberry Pi boards and other sensors. The results demonstrate the
efficiency of the presented safety scenarios in dealing with the dangers of the road.
1 Introduction
Road safety has become a very important subject during the last years. Every year,
millions of people die and thousands are injured due to road accidents. The World
Health Organization (WHO) reported that 1.25 million people died in 2013 and
more than one million accidents were recorded in 2014 in Europe, resulting in 1.4
million injuries and 25 900 deaths. In Morocco, 3593 persons were killed in 2016
against 3499 in 2017 [1], which represents a limited reduction of 2.6% year to
the other. Even if the statistics show a minor decrease, a lot of worrying facts are
hidden [2]. These statistics are so flagrant that the Ministry of Equipment, Transport,
Logistics, and Water (METLE) dedicated a call for projects especially in the field
of road safety, in cooperation with the National Center for Scientific and Technical
Research (CNRST) and Moroccan universities. The aim of those projects is first
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1101
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_101
1102 Z. Sabir and A. Amine
of all decreasing the number of accidents that occur every year in Morocco. Our
project entitled “SafeRoad: Multiplatform for Road Safety (MRS)” is one of those.
His main goal is to reduce the collision rate using the anticipation and the detection
of accidents. And since the collection of statistics of the corporal accidents of road
traffic carried out by the METLE in 2016 [3] reported that there were 80680 accidents
caused by obstacles, we considered also working on the detection of obstacles using
vehicular communications.
The automation/semi-automation of vehicles can help humans avoiding errors
that cause most road accidents. Constructors are developing smart vehicles that will
be supplied by a variety of technologies and connected. They will assist drivers
by alerts indicating the dangers on the road. Principally, they will be connected to
a variety of sensors within the vehicle: warning lights, stability, diagnostic tools,
airbags, and others. Thus, a vehicle is able to automatically disseminate information
to others, like the presence of a slippery roadway, if it is down or in an accident. If the
technologies of networking, electronics, and computing are well integrated, the roads
will be safer and the impact on the environment will be minimized [4]. Recently,
diverse definitions have been given to a “connected” vehicle. A description that
involves all other connections of a smart vehicle is provided by the US Department
of Transportation (USDOT): “Connected vehicle applications provide connectivity
between vehicles to prevent accidents, between vehicles and infrastructure, in order to
generate safety, mobility and environmental benefits; between vehicles, infrastructure
and wireless devices to provide continuous real-time connectivity to all users of the
system” [5]. The system of connected vehicles [6] will operate using Vehicle to
Infrastructure (V2I) [7] and Vehicle to Vehicle (V2V) [8] communications.
In this paper, we present our solution to the problem of road accidents, especially
those caused by obstacles. We introduce an implementation of a vehicular commu-
nications testbed. We focus on V2V technology, and we carry out our experiments
using mini cars based on Raspberry Pi boards. We add different sensors and features
to develop our complete prototypes and we propose two road safety scenarios that
can help to avoid accidents. The next sections of the paper will be as follow: Sect. 2
discusses the related work, Sect. 3 recalls some vehicular network characteristics.
The proposed system is presented in Sect. 4, Sect. 5 Concludes the paper.
2 Related Work
• FleetNet: This German project was initiated by six companies and several research
institutes [12]. The project aimed to develop and demonstrate a wireless ad-
hoc network for vehicular communications through a platform of cooperative
driver-assistance systems, in order to ensure comfort and security for drivers and
passengers.
• The Car2Car communication consortium was inaugurated by European automo-
bile manufacturers. The goal was to improve road safety and to manage effectively
the traffic thanks to the use of inter-vehicle communications [13]. This project had
several missions and intends to develop a V2V system prototype for road safety
applications.
• CarTalk2000 [14] is a European project which aimed to establish, test, and evaluate
an organized and pure ad hoc wireless network and also to develop cooperative
driver assistance systems, in order to enhance global traffic safety and to increase
driving comfort and transport efficiency.
• NoW (Network on Wheels) [15] is a German research project supported by the
government and implemented by research institutes and universities, providers,
and vehicle constructors. The goal is to develop a vehicular communication system
dedicated to road safety, traffic efficiency and infotainment applications, radio,
security, and privacy.
• The Integrated Project PReVENT [16] is a European automotive industry activity,
started in February 2004 with the goal of establishing and validating preventive
safety applications and technologies. Co-funded by the European Commission,
PReVENT stands for Preventive safety and act in accordance with the objective
of the safety initiative to pull down accidents and traffic death in Europe.
• MobiVIP is a research project of Predit 3 (Integration of the Communication
and Information systems Group) [17]. Its objective is evaluating and showing the
influence of the information and communication technologies on a new service
formulated by small urban vehicles for mobility in town centers.
In our work, different from those researches, we focus on the Moroccan case and
we propose a system that we will be adapted to fit the Moroccan situation since
a foreign system cannot be applied directly in Morocco. Many factors are to take
into consideration such as the culture, the environment, the driver behavior, and the
architecture of the cities. We succeeded in the first step using mini cars and we will
move to real tests with real vehicles in the next step.
features and restrictions related to vehicular networks which differentiate them from
ad hoc networks:
• Neighborhood recognition: drivers can receive various information that will
help them ameliorate their visibility and behave accordingly to any adjustments
in the surroundings of the vehicle. This information could be the state of the road,
weather conditions… etc.
• Elevated processing capacity: vehicles can provide powerful communication,
sensing, and computing capacity. Nowadays, they are capable of analyzing and
interpreting the assembled data in order to help drivers to make good decisions.
• Extensive scope: as opposed to other ad hoc networks that adopt generally a
restricted network scope, vehicular networks are able to reach the whole route
system and thus cover a lot of users.
• Unlimited transmission power and adequate storage: vehicles have big storage
and can provide stable power to computing and communication equipment. They
can have also different communication interfaces such as Bluetooth, Wi-Fi, and
others.
• Changes of the topology: vehicular networks are distinguished by fast move-
ments due to the vehicles’ speed that can reach 200 km/h. Therefore, nodes connect
and disconnect quickly and the topology changes frequently. This points to poor
connectivity and minimal road life.
• Anticipated movement: usually, vehicles’ mobility is restricted by the roadways,
they are often connected to Road Side Units (RSUs), roadside infrastructures
(like panels and traffic lights), or highways. This makes the anticipation of their
movement easier, in contrast to classic mobile ad hoc networks, where nodes
move arbitrarily.
• Diverse communication status: the two habitual communication status in vehic-
ular networks are outside and inside the town. In the first one, the area is almost
easy and the roads are usually straight (especially on highways). However, in the
second one, the circumstances are quite difficult. Inside the town, obstacles such
as buildings may isolate the paths. Consequently, communication is not always
continuous.
4 Proposed System
Today’s traveling manner can be transformed and improved through the use of
connected vehicles. They will help to create a wireless communication network, safe
and interoperable, which includes various components (trains, cars, buses, traffic
light …). Connected vehicle technologies address some huge challenges in several
fields such as environment, safety, transportation, and mobility.
The aim of our work is part of the road safety field, particularly in the Moroccan
case in order to resolve the problems mentioned above. We propose a new system
which can be used to notify Moroccan drivers in various circumstances. It will
contribute to reducing the number of accidents and avoiding the crash that frequently
A Novel System Based V2V Communications … 1105
occurs between vehicles or between vehicles and bikes, animals, pedestrians, or other
objects. Our system uses V2V communications and it is based on reel time informa-
tion. Vehicles interchange important information that is valuable to predict a risky
situation. Thanks to this system, the danger of unstructured roads in Morocco can
be avoided by delivering the aforementioned information to drivers.
Considering the statistics cited in [3], the first road safety scenario in our system
can alert a driver of an obstacle in the road. An example is depicted in Fig. 1, where
three vehicles are traveling in one lane and the first vehicle stopped because of a cow
crossing the road. The second vehicle is blocking the vision; hence the third vehicle
can’t see the first one. Thanks to V2V communication, the last vehicle is capable of
receiving a warning about the hard braking and avoiding a possible accident.
The second road safety scenario we propose aims to alert a driver about a specific
traffic condition on the road. The driver will then reduce the speed and change
his behavior regarding the situation (accident, road work, …). He may consider
changing the route and taking a safer one. Due to V2V communication, the warning
is transmitted by a vehicle ahead that has already discovered the problem. Figure 2
shows an explicative scheme of our system.
test zone using a local ad hoc network. They are also equipped with a step-down
DCDC converter module, a USB camera, and a PCA9685-based servo controller.
In order to build our complete prototype, we had to add some other sensors since
the original kit didn’t come with all the necessary equipment. In Fig. 4 we can see the
added sensors which are the infrared obstacle detector and the line follower sensors.
This figure depicts also the location of each sensor. The infrared sensor will be used
A Novel System Based V2V Communications … 1107
to detect any obstacles on the road while the line follower sensors will help to keep
the car on the road. The details of the sensors are provided in Table 1.
Using V2V communications, our road safety scenarios aim to disseminate an alert
informing the drivers of the existence of an obstacle on the roadway. As a first try,
we attempt to run the cars on a map, so they follow pre-programmed distances.
The result was that the cars couldn’t move following a straight line, due to the
little difference between the motors of each car. Therefore, we preferred to use line
follower sensors to help cars following the road. The first road safety scenario is
about the hard-braking warning. We put an obstacle on the road and ran the two
cars at once. Since the first car (V1) is equipped with an infrared sensor (see Fig. 4),
it detects the obstacle and sends the alert to the second car (V2) using the Wi-Fi
interface. The moment it received the alert, V2 stops as well (see Fig. 5). Receiving
the warning message depends on the Wi-Fi range (the maximum distance between
V1 and V2).
The second road safety scenario aims to send a warning to the drivers, so they
can choose the safer road. Similarly, we put an obstacle on the path before the final
destination and we ran V1 and V2 at the same time.
Once the infrared sensor senses the obstacle, V1 stops instantly and sends the
alert to V2 which changes the current path (taken by V1) and chooses the safer one.
Figure 6 illustrates the second road safety scenario.
We present a comparison of our proposed system with other studies in the literature
in Table 2.
5 Conclusion
We reached our target and succeeded in achieving the solution. The primary consider-
ation of the paper was to introduce an implementation of a complete vehicle commu-
nication system using V2V technology in the context of the SAFEROAD project.
Thanks to the partnership with the METEL, we obtain real information about road
accidents in Morocco. Our solution is developed especially for the Moroccan case
and it is the first of its kind in our country. The aim was to send an alert to the drivers
in time, in order to help them avoiding road dangers and so decreasing the number
of accidents. We conducted our experiments using mini cars based on Raspberry Pi
as well as infrared obstacle detector and line follower sensors. The results affirm the
usefulness of our system in the field of road safety. In future work, we will confirm
the reality on the ground by applying this system to real cars and test it in different
cities, especially those where a grand number of accidents occur.
Acknowledgment This research work is supported by the “SafeRoad: Multiplatform for Road
Safety (MRS)” Project under contract No: 24/2017, financed by the Ministry of Equipment, Trans-
port, Logistics and Water (METLE), and the National Center for Scientific and Technical Research
(CNRST).
References
1. World Health Organization (2018) Global status report on road safety 2018 (WHO). World
Health Organization, Geneva
2. Leconomiste: Accidents routiers. https://www.leconomiste.com/flash-infos/accidents-routiers-
au-maroc-tue-presque-autant-qu-en-france-avec-7-fois-moins-de. Accessed 08 July 2019
3. Ministry of Equipment, Transport, Logistics, and Water, Direction of Roads (2017)
Compendium of Statistics of Road Traffic in Personal Accidents
4. Guerrero-Ibáñez JA, Flores-Cortés C, Zeadally S (2013) Vehicular ad-hoc networks (VANETs):
architecture, protocols, and applications. In: Chilamkurti N, Zeadally S, Chaouchi H (eds)
Next-generation wireless technologies. Springer, London, pp 49–70
5. Brookes R, Pagani P (2014) What becomes a car. In: Proposed paper for: BIT 2014 conference
workshop-technology enabled business models: platforms, analytics and performance, March
2014
6. Sabir Z, Amine A (2020) NDN vs TCP/IP: which one is the best suitable for connected vehi-
cles? In: Proceedings of the first international conference on technology, engineering, and
mathematics, Kenitra, Morocco, 26–27 March 2018. Springer, Cham, pp 151–159
7. Sabir Z, Dafrallah S, Amine A (2019) A novel solution to prevent accidents using V2I in
Moroccan smart cities. In: 2019 international conference on computational intelligence and
knowledge economy (ICCIKE), Dubai, UAE, pp 621–625
1110 Z. Sabir and A. Amine
8. Sabir Z, Amine A (2018) Connected vehicles using NDN for intelligent transportation systems.
In: International conference on industrial engineering and operations management, Paris,
France, pp 2433–2441
9. Tsugawa S (2005) Issues and recent trends in vehicle safety communication systems. IATSS
Res 29:7–15
10. Hedrick JK, Tomizuka M, Varaiya P (1994) Control issues in automated highway systems.
IEEE Control Syst Mag 14:21–32
11. Gehring O, Fritz H (1997) Practical results of a longitudinal control concept for truck
platooning with vehicle to vehicle communication. In: Proceedings of conference on intelligent
transportation systems, Boston, MA, USA. IEEE, pp 117–122
12. Enkelmann W (2003) FleetNet - applications for inter-vehicle communication. In: IEEE IV2003
intelligent vehicles symposium. Proceedings (Cat. No. 03TH8683), Columbus, OH, USA.
IEEE, pp 162–167
13. C2C-CC: CAR 2 CAR Communication Consortium. https://www.car-2-car.org/. Accessed 24
June 2019
14. Reichardt D, Miglietta M, Moretti L, Morsink P, Schulz W (2002) CarTALK 2000: safe and
comfortable driving based upon inter-vehicle-communication. In: IEEE intelligent vehicle
symposium, 2002, vol 2, pp 545–550
15. Festag A, Noecker G, Strassberger M, Lübke A, Bochow B, Torrent-moreno M, Schnaufer
S, Eigner R, Catrinescu C, Kunisch J (2008) ‘NoW – network on wheels’: project objectives,
technology and achievements, Hamburg, Germany
16. Schulze M, Nöcker G, Böhm K (2005) PReVENT: a European program to improve active
safety. In: Proceedings of 5th international conference on intelligent transportation systems
telecommunications, France
17. Despeyroux T, Masseglia F, Rossi F, Senach B (1992) User-centered design, improvement and
analysis of information systems. Bull Sociol Methodol/Bulletin de Méthodologie Sociologique
37:55–57
18. Sunfounder: Smart Video Car for Raspberry Pi. https://www.sunfounder.com/learn/category/
Smart-Video-Car-for-Raspberry-Pi.html. Accessed 15 July 2019
19. Buy a Raspberry Pi 3 Model B+ – Raspberry Pi. https://www.raspberrypi.org/products/raspbe
rry-pi-3-model-b-plus/. Accessed 01 Aug 2019
20. Ghatwai NG, Harpale VK, Kale M (2016) Vehicle to vehicle communication for crash avoid-
ance system. In: 2016 international conference on computing communication control and
automation (ICCUBEA), pp 1–3
21. Hadiwardoyo SA, Patra S, Calafate CT, Cano J, Manzoni P (2017) An android ITS driving safety
application based on vehicle-to-vehicle (V2V) communications. In: 2017 26th international
conference on computer communication and networks (ICCCN), pp 1–6
22. Tayeb S, Pirouz M, Latifi S (2017) A raspberry-pi prototype of smart transportation. In: 2017
25th international conference on systems engineering (ICSEng), pp 176–182
Detection of Underground Cavities Using
Electromagnetic GPR Method in Bhalil
City (Morocco)
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1111
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_102
1112 O. Jabrane et al.
1 Introduction
Karst processes often result in the formation of underground cavities due to the effect
of chemical erosion by the circulation of acidic water (dissolution) on carbonate
rocks.
The urban area of Bhalil continues to develop locally on more or less unstable
grounds, in particular its old town, subjected to multiple constraints. These are linked
to rapid population growth coupled with the fragile geological nature of the under-
lying formations. Another critical aspect of these cavities lies in the fact that they
can provide a safe and consistent habitat for many people in this area.
Underground cavities detection is one of the primary objectives of geophysical
investigation techniques because of the danger that threatens the inhabitants causing
soil instability, which can provoke dramatic events such as subsidence and even total
collapse to the ground surface [6, 15]. The diagnosis of these buildings condition is
carried out regularly by the authorities sometimes forces them to evacuate certain
houses deemed to be in danger. For the subsurface identification and mapping of
such karst features, the non-invasive and high-resolution geophysical techniques
have appeared as an appropriate choice [12].
The present case study, applied Ground-penetrating radar (GPR) for the site char-
acterization, which uses radar pulses to image the subsurface. It is a non-intrusive
prospecting technique based on the propagation of electromagnetic waves in the
underground to explore the presence of a medium RDP [13]. It has become one of
the most important geophysical methods for cavities and voids detection, to delimit
them laterally and in-depth according to its efficiency, high resolution, rapid execu-
tion and simultaneous result by imaging the subsurface. Mapping dangerous areas
in the Old Medina are an asset in terms of future urban development in the city,
protection of people’s properties, and especially their lives.
Our investigation consists of two complementary parts: i/ Geological study to
identify the origin of the observed voids and their consequence at the surface; ii/
prospecting and locating the underground cavities using indirect and non-destructive
electromagnetic GPR method.
The study area is located in a small parcel around the old town of Bhalil, which is
28 km south of the city of Fez. The town of Bhalil is 950 m above sea level, it is a
transition zone between the tabular Middle Atlas, to the south, and the Sais basin, to
the north.
Detection of Underground Cavities Using Electromagnetic GPR Method … 1113
Karst features presence greatly increases the risk of underground cavities devel-
opment by a cluster of inter-related processes, including bedrock dissolution, soil
sapping, rock collapse, and soil collapse. The presence of underground cavities in
the basement of an urban area can cause severe effects and threatens the lives of the
inhabitants.
Fig. 1 A-Geological section showing the layout of the outcrops in the study area [7], B-
Lithostratigraphic column showing the main formations described in the study area
1114 O. Jabrane et al.
(villafranchien) tufa and travertine deposits. The latter, contains many voids and
cavities, due to the natural karst phenomenon, or to the human action consisting in
digging into the rocks for underground habitats. A final recent layer consists of late
quaternary (colluvial deposits that conceal the cavities).
3 Methodology
This study was conducted to detect subsurface cavities using GPR (Ground Pene-
trating Radar) designed by the Swedish company Mala geosciences. The advantage
of this technique is its high-resolution results, faster operation, equipment portability
and simultaneous picturing of sub-soil in terms of RDP.
This technique is based on the use of two antennas, one for transmitting electro-
magnetic (EM) waves, and the other for receiving these waves after reflection on
the RDP targets. The two antennas are connected to a console, itself connected to a
laptop that allows data visualization during acquisition, or to configure the console. A
simple automotive battery supplies the power, the transmitter sends electromagnetic
waves, ranging between two frequencies from 10 MHz to 2 GHz. The signal is not a
single sinusoid at the designated frequency, but a sum of sinusoids whose spectrum
is centered on that frequency [2].
The investigation depth range using GPR method depends on the DPR which is
related to the electrical conductivity of the underground layers, transmitted center
frequency of antennas and the radiated power [8, 10], while the EM velocity is
conditioned by the dielectric permittivity of the soils [1, 5, 14].
The dielectric permittivity can be expressed as:
ε = ε = εr ε0 (1)
1/ 2
με σ 2 1/ 2
∝= ω 1+ −1 (2)
2 ωε
⎡ ⎤−1/2
σ 2 1/2
2 ⎣
V = 1+ − 1⎦ (3)
με ωε
When σ is small as compared with ωε, the EM velocity can be expressed as:
C
V =√ (4)
εr
The measurements were carried out in a parcel of the Rquiba district (Fig. 2), using
two antennas RAMAC GPR. For all ground-penetrating radar profiles, it was neces-
sary to choose some characteristics of implementation in the field. After several tests
to avoid sever attenuation and to get the best resolution, we opted for a recording
Fig. 2 A Satellite image showing the survey area, B map showing profiles arrangement in Rquiba
district
1116 O. Jabrane et al.
time of 200 ns. It corresponds, in the context of the study area, to an investigation
depth of about 10 m and bistatic type antennas with a central frequency of 200 MHz.
The profiles have different lengths and different orientations, 6 profiles which are
oriented NNE-SSW perpendicular to the length of the recognized cavities in the
studied area, separated by a distance of 1.2 m, and 4 profiles perpendicular to the
previous profiles separated by a distance of 3 m.
3.3 Processing
The geophysical ground penetrating radar work carried out as part of this study
enabled us to record a large amount of data. The radargrams collected systematically
underwent the processing sequence.
To improve the image provided by the radiograms, it is necessary to apply certain
processing to the raw data from the ground penetrating radar, in order to increase
the signal/noise ratio of the radar sections, and to make reading and interpretation
easier. They help to attenuate the physical phenomena inherent in the propagation of
electromagnetic waves in the ground so as to obtain an image that reflects the object
to be identified. The processing applied in our work include in particular:
• Time-zero correction for shifting the traces along the time axis.
• Removing the low-frequency components (“wow” operation) of the GPR record;
it reduces the data to a mean zero level.
• Band-pass filtering (1D filter band pass frequency): Eliminate the high-frequency
(noise).
• Signal gain «energy decay»: Applying a gain to compensate the signal attenuation.
The final sections obtained were then carefully analyzed in order to extract as
much information as possible on the structure of the subsoil investigated and on the
possible existence of underground cavities within the plot.
The ground-penetrating radar profile shows areas of strong linear reflections and
areas of very low amplitude. We associate these with the voids contained in the
travertines layer, numbered A1, A2, and A3 (Fig. 3). We can follow spectacularly
the first decimeters of travertines formation before encountering disturbances linked
in our case to the presence of 3 anthropic cavities.
Alongside the anomalies related to voids, discontinuities have been highlighted
that we associate to small fractures, which are sloping towards the North East.
Strong reflections continue up to 20 ns. We measured the thickness of the roof
of the cavities related to a small subsurface travertine formation, which was about
0.5 m on average.
Detection of Underground Cavities Using Electromagnetic GPR Method … 1117
Fig. 3 GPR Profile 1 with depth scale, showing the arrangement of the 3 cavities
Fig. 4 GPR Profiles (a, b, c, d. and e.) showing all A1 anomalies, (a, and b. profiles) showing the
second cavity anomaly A2
1118 O. Jabrane et al.
Fig. 5 Perpendicular GPR Profiles (f, j, h, and i.), radar-gram (f, and h.) showing the same anomaly
A2
We can follow these voids in profiles 3, and 4 (Fig. 4), in which we note the
absence of anomaly A3. As a result, the cavity linked to this anomaly is very limited
in space. The cavities corresponding to anomalies A1 and A2 continue deeper to
appear in profile 4, implying a length of about 6 m. From the fifth profile, we note
that all kind of anomalies disappear, so that indicates a more or less homogeneous
level of the same travertine formation.
The profiles from 7 to 10 (Fig. 5), are perpendicular to the first ones, so they follow
the cavities but present little disturbance. The profile 7 and 9 which are perpendicular
to profile 1 and passes through the middle, show an anomaly appears close to the
surface, attributing in our case to anomaly A2 of profile 1 which indicates the presence
of an underground cavity in the same travertine formation.
5 Conclusion
In this work we demonstrated the efficiency of the GPR application as a very powerful
tool for mapping distribution of different anomalies related to cavities in urban area.
Detection of Underground Cavities Using Electromagnetic GPR Method … 1119
The results are very encouraging, as most of the cavities reported by the inves-
tigation have been confirmed. We were able to follow these voids limit in the three
directions of space at the scale of the plot studied. It was therefore possible to select
the largest cavities from the small ones thanks to the large number of profiles and
their arrangements. The very fine spacing of the measurements (20 cm) also made
it possible to highlight small fractures that affect the travertines, which are very
limited in space (a few meters) and sloping NE direction, which is consistent with
the relatively large structures mentioned in previous geological work in the area.
In this work the description of danger zones (areas with potential risk evidenced
by cracks and voids in-depth), requires additional GPR surveys at the western central
limit of the urban area in the old Bhalil, to define the necessary preventive measures.
A better geotechnical knowledge of the underlying layers would also be beneficial
seeking to minimize the risk of the possible collapse and landslides. From a geotech-
nical point of view, all the new infrastructure and facilities must be located in a stable
area to avoid construction disorder which represents a high financial impact for many
owners.
Funding This research was funded by AECI project: A/012172/07: Geophysical investigating for
the delimitation of the unstable zones of the city of Fez and vicinity.
References
1. Annan AP (2003) Ground-penetrating radar principles, procedures & applications. Sensors and
Software Inc, Mississauga
2. Bano M (1996) Constant dielectric losses of ground-penetrating radar waves. Geophys J Int
124(1):279–288
3. Charriere A (1990) Héritage hercynien et évolution géodynamique alpine d’une chaîne intra-
continentale: le Moyen Atlas au SE de Fès (Maroc). Thèse, Doctorat d’Etat, Toulouse, p
589
4. Charroud M, Cherai B, Benabdelhadi M, Charroud A, El Moutaouakkil N, Falguères C, Lahrach
A (2006) Sedimentary evolution of a fore-chain Sais basin during plio-quatrenary and modali-
ties of tectonic inversion (Sais basin, Morocco). In: Geophysical research abstracts, European
geosciences union, vol 8, p 10039
5. Daniels DJ (2004) Ground penetrating radar. The Institution of Electrical Engineers, 2nd edn.
London
6. De Bruyn IA, Bell FG (2001) The occurrence of sinkholes and subsidence depressions in the far
West Rand and Gauteng province, South Africa, and their engineering implications. J Environ
Eng Geosci 7(3):281–295
7. Fassi D (1999) Les formations superficielles du Saïs de Fès et de Meknès. Du temps géologique
à l’utilisation actuelle des sols. Notes et Mémoires du Service Géologique, Maroc, pp 389–527
8. Franke JC, Yelf RJ (2003) Applications of GPR to surface mining. In: Proceedings of the 2nd
international workshop on advanced ground penetrating radar, University of Technology, Delft,
The Netherlands, pp 115–119
9. Gourari L, Boushaba A, Ramon J, Akdim B (2000) Hydrochimie, processus, facteurs de
précipitation des encrôutements travertineux et les causes géo-environnementales du ralen-
tissement de leur formation actuelle dans la bassin karstique de l’Oued Aggai (Causse de
Sefrou, Moyen-Atlas, Maroc). Comm Inst 115–133
1120 O. Jabrane et al.
Ali Ouacha
Abstract In the current time, technological progress produces devices that have
become smaller in size. Performing tasks by these devices is in most cases a deli-
cate task due to their limited performances (such as processing capacity, storage and
connectivity), especially when dealing with large data size and when having time
constraints. However, Mobile Edge Computing (MEC) comes to overcome these
weaknesses by delegating the execution of heavy tasks to other peripheral nodes
with more powerful in computing capacity (Edge Server). In addition, these devices
form a mobile network in which communication between them is based on routing
protocols. Thus, in this paper, we take benefits from the control information
exchanged between different nodes of the network to dissimulate information
concerning the processing capacities of nodes, and to take advantage of physical
links characteristics between nodes. All these information is combined into a metric
allowing the selection of the most privileged neighbor node to perform various tasks
of a particular node. Intensive simulations are done by the NS-3 simulator showing
the effectiveness of this idea especially in runtime.
1 Introduction
Technological progress have led to the birth of smaller equipment that generates large
amount of data. This has created the need to work together and collaborate to achieve
a number of goals. Indeed, in an environment made up of several devices, some of
them have very limited computational performance. To be able to process a set of
tasks within acceptable deadlines, it must rely them on another powerful equipment.
It is the key concepts of Mobile Edge Computing (MEC) [1, 2]. In addition, in order
A. Ouacha (B)
Computer Science Department,
Faculty of Sciences, Mohammed V University, Rabat, Morocco
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1121
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_103
1122 A. Ouacha
to communicate, all these devices use a routing protocol. One of these protocols and
most commonly used is the Optimized Link State Routing Proto-col (OLSR) [3, 4]
It is classified under the proactive protocols category whose principle is based on the
periodic exchange of information on the state of the links and on the neighborhood
through several types of control messages. Thus, so that each node can learn about
the performance of the other nodes of the network, several types of information are
added to the exchanged control messages. So, the main objective of this work is to use
these information that present the computation characteristics of each neighboring
node and combine them with other information extracted from the characteristics
of the physical link that connects the concerned node to other neighboring nodes
in order to be able to choose the best node that is capable of perforing a task list.
Moreover, since this is a mobile environment, the time required to process the tasks
by a neighbor node must be taken into consideration. This is what the predicted time
to quit the neighborhood is used for. It allows the node to avoid selecting a neighbor
node which risks leaving the neighborhood without even finishing execution of tasks
assigned to it.
The whole article has several sections in addition to the introduction and
the conclusion. The first section illustrates an overview of the OLSR protocol.
The second section provides a state of art that summarizes the various works on
the concept of treatment sharing. The third section presents the formulation of the
problem as well as the proposed solution whereas the fourth section presents the
various tests and simulations carried out as well as the obtained results.
2 OLSR Overview
3 Related Works
The authors of the research made in [6] present an algorithm based on a task
scheduling approach that is founded on dynamic voltage and frequency scaling
(DVFS). It aims to achieve energy reduction by allocating tasks to the appropriate
cores of mobile devices or offloaded it to the mobile cloud computing (MCC) envi-
ronment. The authors of the research made in [7] propose an heuristic based approach
named Edge Cloud Heuristic Offloading (ECHO) to find a tradeoff solution between
application runtime, battery lifetime, and user cost according to the user’s prefer-
ences. In the research made in [8], the authors propose a solution based on game
theory for offloading tasks among a swarm of Unmanned Aerial Vehicles (UAVs).
They opt to reduce their energy consumption and guarantee QoS for users by consid-
ering the UAVs as aerial capacitated cloudlets pooling their computational resources
to execute tasks from the ground internet of things (IoT) devices in their coverage.
The authors of [9] propose an online energy efficient task assignment and computing
offloading strategy in the aim to minimize the time average energy consumption of
the multimedia services. The proposed strategy can adaptively decide the task assign-
ment, coordinate and optimize the wireless and computation resource allocation by
taking the dynamic wireless condition and service delay constraints into considera-
tion. In [10] authors propose a framework to provide guaranteed services at minimum
system cost. To address the challenges of configuring the edge optimization capa-
bility and determining the type and amount of clouds in which resource provision
within the Cloud Assisted Mobile Edge (CAME) framework, the authors provide
an Optimal Resource Provisioning (ORP) algorithm that is used for optimization of
the capacity of adjacent hosts and provides dynamic cloud renting strategies. The
authors of the research made in [11] propose a model that uses the Markov approx-
imation method for full implementation of tasks and computations in MEC. This
model provides a multi-server system in which only one mobile is brought close to
1124 A. Ouacha
the neighboring servers for complete computing. The use of the Markov approxi-
mation technique provides a common optimization of the allocation of computing
tasks and the frequency of the processor’s scalability to minimize the runtime and
mobile power consumption. The research made in [12] proposes a new MEC archi-
tecture based on intelligent computational offloading. Depending on the size of the
data from the mobile user computing task and the performance characteristics of
the periphery computing nodes, a task-predicting and task migration algorithm are
proposed. In the work done in [13], the authors propose an Energy-aware Collabora-
tive Routing (ECoR) scheme for optimally handling task offloading between source
and destination UAVs in a grid-locked UAV swarm. The proposed ECoR acts on two
objectives: routing path discovery and routing path selection. In the first one (routing
path discovery), the scheme selects the most optimal path based on the maximization
of residual energy. In the second one (routing path selection), the scheme ensures
balanced energy utilization between members of the UAV swarm and enhances the
overall path lifetime without incurring additional delays in doing so.
Like these researches, our goal is to find for a mobile node (MN), whose compu-
tational performance is limited, the best neighbor node (NN) that presents the ideal
environments to perform its tasks. But unlike all these works, we seek taking advan-
tage of the normal operating of the OLSR protocol and insert in the control messages
exchanged between nodes the characteristics relating to the processing capacity.
As shown in Fig. 2, the Mobile Node (MN) contains a list of N independent tasks,
where each one is denoted τi with i = {1, 2, . . . , N }. Those tasks that are supposed
to be computationally greedy and sensitive to delays should all be achieved. The
execution time of the task list cannot exceed a maximum required time t max and each
task τi can be processed remotely by an elected Neighbor Node (NN).
Each data task is atomic (i.e. it cannot be divided into subtasks) and it is charac-
terized by the following two parameters τi di , λi . The first one denoted di [bits]
represents the size of the task (quantity of input parameters and program codes) to
be transferred from the MN in question to the NN. The second parameter, denoted
λi [cycles], specifies the workload referring to the amount of computation necessary
to accomplish the processing of this task. In addition, the execution time of a task τi
also depends on the frequency f j [H er t z] of the N N j . Considering the constraint of
a mobile environment, the NN should be chosen from nodes in the first neighboring
so that it stays in the neighboring of the MN as long as possible.
and for all the tasks to be executed on the N N j , we have:
An Adapted Routing Protocol for Mobile Edge Computing 1125
N λi
t exec
j = [Second] (1)
i=1 fj
Since the task τi is offloaded to the N N j , the time needed to transmit it is given
by:
di
ti,comm
j = [second] (2)
rj
In this equation r j corresponds to the uplink data rate [14]. It is calculated by:
PT × gj
r j = w × log(1 + ) (3)
σj
With: w is the bandwidth of the link between the MN and N N j , note that the
MN and all nodes of the neighborhood use the same channel. P T represents the
transmission power of the MN, σ j the noise of the channel to which the MN and the
N N j are connected and g j the gain between the MN and the N N j .
and for all the tasks to be offloaded on the neighbor node N N j , we have:
N di
t comm
j = [Second] (4)
i=1 r j
1126 A. Ouacha
Thus, the completion time of the process of offloading and executing a single task
τi , which will be called Time Metric (TimeMetric) is given by:
di λi
ti, j = + [Second] (6)
rj fj
In this section, we present our formulation of the optimization problem that aims at
selecting the convenient neighbor node N N j whitch minimizes the time t j it takes
for an offloaded task set to be executed. In other words, this is to select the node
corresponds to the lowest value t j from the possible set of available values
that
T t1 , t2 , . . . , tj , . . . , tM and that minimizes our TimeMetric.
The obtained problem is formulated as follows:
⎧
⎪
⎪ ArgMin {T }
⎪
⎪ { f j ,r j ,λi ,di }
⎨
P1 (C1 ) f j > F min (8)
⎪
⎪
⎪
⎪ (C2 ) t j > T min
⎩
(C3 ) e M N > E min
threshold E min in order to avoid the nodes that risk shutting down before achieving
the execution of the tasks. The threshold E min for a single task is calculated by
the Eq. 10 below.
P T λi
E imin = [J ] (9)
fj
N P T λi
E min = [J ] (10)
i=1 fj
In the formulation part of the problem, Constraint (C2) specifies that the execution
time of all the offloaded tasks must be less than threshold T min in order to avoid
choosing the NN which risks leaving the neighbourhood of the MN because of
mobility constraint. The time T min in question will be represented by the predicted
Rest Time to Quit (RTTQ) calculated as illustrated in Fig. 3 [14]. It is estimated based
on the cartesian coordinates exchanged through the HELLO messages received from
NN.
The TimeMetric is calculated when the MN has a list of tasks to transmit to be
executed on the corresponding neighbor node N N j using algorithm 1. Note that the
frequency f j , λi as well as the node coordinates x j and y j of each neighbor node
are retrieved from the HELLO message sent and they are stored by the MN.
Wherever Times is specified, Times Roman of Times New Roman may be used. If
neither is available on your word processor, please use the font closest in appearance
to Times. Avoid using bit-mapped fonts if possible. True-Type 1 fonts are preferred.
5.1 Simulations
To test the solution presented in this work, we used as a simulator NS-3. It has the
advantage of being more suitable for network environments. It contains an imple-
mentation of a very large number of protocols, most often used in various types of
networks. In addition, it is a free and open source. The network architecture used in
NS-3 is based on existing software and hardware in the real world. This facilitates
the use of the simulator and allows easy integration of new models.
The simulation environment as well as the simulation parameters, as shown in
Table 1, consists of a network of 50 mobile nodes according to the Random Waypoint
Mobility model [15]. All nodes have a single interface that uses OLSR as the routing
protocol. A variable number of nodes between 5 and 25 with a step of 5 are chosen
to send a set of tasks represented by a data packet of size 1000 Bytes. Each node
sends 10 packets with a frequency of 1 packet for one (1) second.
An Adapted Routing Protocol for Mobile Edge Computing 1129
Table 1 Simulation
Parameter Value
parameters
Network simulator NS −3.29
Simulation time 100 s
Simulation area 1500 × 1500 m × m
Number of nodes 50
Transmission range 500 m
Mobility model Random WayPoint (RWP)
Max speed 20 m/s
Max pause time 0s
Packet size 1000 bytes
WiFi 802.11b
To simulate the calculations performance, two numbers are assigned to each node.
The first one represents the frequency f j of the N N j processor. It is initialized by
a randomly generated value between 0.5 and 3.0 GHz. The second represents the
number of cycles λi necessary to process a single task. It is initialized by an integer
value generated randomly between 1 and 8.
The simulations consist in choosing a number of source nodes; each one generates
10 data packets constituting the tasks to be executed. Before sending each set of
tasks, each node researches through the nodes of its first neighborhood for the one
that has the most minimal possible TimeMetric. Given the mobile environment and
to avoid the choice of a neighboring node which risks leaving the neighborhood even
before giving the result of the processing, only the one whose RTTQ (algorithm 1)
is twice higher than the TimeMetric is chosen. In order to be able to calculate the
communication time which is part of the TimeMetric, the information that will be
needed is extracted from the physical layer of MN as well as from the frame received
from the NN containing the hello message.
5.2 Results
At the end of the simulations under NS-3, a set of results are obtained. Thus, the
graph of Fig. 4 clearly shows that the average time necessary for the execution of the
tasks by the selected NN always remains lower than the average time calculated at
the level of all the other neighboring nodes necessary for the execution of the same
set of tasks regardless of the number of source nodes.
In addition, Fig. 5a shows that the TimeMetric has experienced a slight increase
when the number of source nodes increases. Indeed, once the number of source nodes
increases, the number of generated packets also increases which causes pressure on
the elected NN. Thus several packets will be directed to this node which generates
waiting times whether it is at the level of the reception of the packets by the lower
1130 A. Ouacha
0
5 10 15 20 25
Number of Source Nodes
Fig. 4 Average time required for the execution of tasks by a neighbor node compared to the average
time calculated at the level of other neighboring nodes
0.66 2.18
0.64 2.16
0.62 2.14
0.6 2.12
0.58 2.1
5 10 15 20 25 5 10 15 20 25
Number of Source Nodes Number of Source Nodes
a. Selected Sink node b. All Neighbor Nodes
layer or at the level of the order of execution of the tasks on the microprocessor.
Figure 5b also shows a slight decrease in the average of the calculated time metric
at all neighboring nodes as the number of source nodes increases.
Despite the gain recorded in terms of the TimeMetric, this system has the disad-
vantage of an additional load that will be added to the control traffic generated by the
OLSR protocol itself. Indeed, Fig. 6 shows a slight increase in the average throughput
exhibited by the Modified OLSR version compared to the standard OLSR version.
This is due to the fact that the HELLO control messages sent every two seconds are
enriched by other information used for the calculation of the time metric and the
RTTQ (current position of the sending node of the HELLO message, its calculation
frequency, the number of cycles of its processor, …).
An Adapted Routing Protocol for Mobile Edge Computing 1131
0.5
0
5 10 15 20 25
Number of Source Nodes
Fig. 6 Average flow rate comparison between the standard version of OLSR Protocol and the
modified version
6 Conclusion
This article illustrates the case of a node whose computational performance is very
limited (the case of a smart watch or a smart garment …) and that has a set of tasks to
perform. To do this, it seeks among the nodes from the first neighbourhood the one
which presents the smallest metric of time. Thus, a first solution consists in taking
advantage of the control messages of the OLSR protocol to exchange the calculation
performances of each node of the network. So each node has everything it needs to
decide which node will run this task list.
The results of the simulations carried out by NS3 show a gain in terms of the time
metric against a price paid by a slight increase in the average flow. These results also
show that (Disadvantages) in a cell, the neighbour node which has the smallest time
metric will be elected by all the other neighbouring nodes. This presents a resulting
pressure by sending multiple sets of spots to that same node.
As Perspectives, a forthcoming study aims to integrate a mechanism to balance
the loads so as to share the processing of tasks between several neighbouring nodes.
A first proposition consists in finding the ideal node or nodes for the execution of the
tasks in the set of nodes of the first and second neighbourhood. Another perspective
this time is to also take advantage of other metrics such as energy performance and
those of the main memory in the search for the execution node.
References
1. Cao J, Zhang Q, Shi W (2018) Edge Computing: A Primer, 1st edn. Springer, Cham
2. Nunna S, Ganesan K (2017) Mobile edge computing. In: Thuemmler C, Bai C (eds) Health 4.0:
how virtualization and big data are revolutionizing healthcare. Springer, Cham, pp 187–203
3. Clausen T, Jacquet P (2003) RFC3626: Optimized Link State Routing Protocol (OLSR), vol
RFC3626: RFC Editor
1132 A. Ouacha
4. Clausen T, Dearlove C, Jacquet P, Herberg U (2014) The Optimized Link State Routing Protocol
Version 2. Internet Engineering Task Force (IETF), April 2014
5. Corson S, Macker J (1999) RFC2501: Mobile Ad hoc Networking (MANET): Routing Protocol
Performance Issues and Evaluation Considerations: RFC Editor
6. Lin X, Wang Y, Xie Q, Pedram M (2015) Task scheduling with dynamic voltage and frequency
scaling for energy minimization in the mobile cloud computing environment. IEEE Trans Serv
Comput 8:175–186
7. De Maio V, Brandic I (2018) First hop mobile offloading of DAG computations. Presented
at the 2018 18th IEEE/ACM international symposium on cluster, cloud and grid computing
(CCGRID), Los Alamitos, CA, USA
8. Ma W, Liu X, Mashayekhy L (2019) A strategic game for task offloading among capacitated
UAV-mounted cloudlets. In: 2019 IEEE international congress on internet of things (ICIOT),
pp 61–68
9. Sun Y, Wei T, Li H, Zhang Y, Wu W (2020) Energy-efficient multimedia task assignment and
computing offloading for mobile edge computing networks. IEEE Access 8:36702–36713
10. Ma X, Wang S, Zhang S, Yang P, Lin C, Shen XS (2019) Cost-efficient resource provisioning
for dynamic requests in cloud assisted mobile edge computing. IEEE Trans Cloud Comput.
https://doi.org/10.1109/TCC.2019.2903240
11. Zhou W, Fang W, Li Y, Yuan B, Li Y, Wang T (2019) Markov approximation for task offloading
and computation scaling in mobile edge computing. Mob Inf Syst 2019:8172698
12. Miao Y, Wu G, Li M, Ghoneim A, Al-Rakhami M, Hossain MS (2020) Intelligent task predic-
tion and computation offloading based on mobile-edge cloud computing. Future Gener Comput
Syst 102:925–931
13. Mukherjee A, Misra S, Chandra VSP, Raghuwanshi NS (2020) ECoR: energy-aware collab-
orative routing for task offload in sustainable UAV swarms. IEEE Trans Sustainable Comput
5(4):514–525
14. Ouacha A, Lakki N, Abbadi JE, Habbani A, Bouamoud B, Elkoutbi M (2013) Reliable MPR
selection based on link lifetime-prediction method. In: 2013 10th IEEE international conference
on networking, sensing and control (ICNSC), pp 11–16
15. Roy RR (2011) Handbook of mobile ad hoc networks for mobility models, 1st edn. Springer,
Heidelberg
Advanced Technologies in Energy
and Electrical Engineering
Improved Geographic Routing Protocol
for Wireless Sensor Networks
1 Introduction
Wireless Sensor Networks (WSNs) [1] are systems made up of several indepen-
dently functioning elements, which are disseminated over large areas. These elements
include limited computing and storage resources (processing units, memories, etc.)
and are fitted with sensors specific to the environments in which they operate and
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1135
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_104
1136 K. Redjimi et al.
to data, which they must measure. They are also equipped with wireless means for
sending and receiving data. The information collected by the various nodes of the
network undergoes a first elementary processing and then is transmitted to the base
station. These sensors can measure various ranges of quantities such as humidity,
temperature, level, pressure, wind speed, objects detection, etc. Consequently, these
sensor networks are useful in several fields and their applications are multiple, for
example in monitoring, control and decision-making. Thus, they are used in medicine
[2], military field, forest or flood fire monitoring, smart homes and cities [3, 4] and
industrial automation among other fields [5].
Batteries supply these elements with energy and thus allow them certain autonomy
of operation. Optimal management of the energetic charge of these batteries remains
one of the most important problems to be resolved with regard to WSNs [6].
A sensor node expends a maximum of energy when it is in the phase of sending
data. The energy expended during the reception is less. However, to optimize the
energy consumption in a sensor node and, consequently, to increase the network
lifetime, it is essential to consider all the energy expenses at all levels of the architec-
ture of the WSN. Indeed, these sensors are sometimes inaccessible and recharging
their batteries can be very restrictive especially when they are located in hostile areas.
As a result, many researchers have carried out the implementation of proto-
cols making it possible to better managing energy consumption in wireless sensor
networks. Note, however, that several solutions provided are in fact improvements
of other pre-existing solutions.
Most of the developed routing protocols for WSNs [7] require node localization
coordinates since events are directly related to the node location. The positions of
two nodes make it possible to calculate the distance between them, which makes it
possible to estimate the energy expended during the communication between these
two nodes. This work exploits this idea to improve the energy consumption in WSN’s.
Our protocol increases packets delivery ratio to the base station.
The following sections are organized as follows: Sect. 2 briefly presents general
notions concerning routing protocols in WSNs. Section 3 is devoted to present the
proposed approach. In Sect. 4; the results and their discussion are presented and
Sect. 5 concludes this article.
2 Backgrounds
the cluster nodes and their routing to the base station. Low-energy adaptive clustering
hierarchy (LEACH) [11] and its variants are typical examples of this class. In the
case of location-based routing protocols, the sensor nodes are accessible thanks to
their positions via the Global Positioning System (GPS), thanks to tables containing
their coordinates or other specific tools. These positions allow the implementation of
information routing. For example, Greedy Perimeter Stateless Routing (GPSR) [12]
is a geographic routing protocol. In Geographic Routing, each node is concerned
only with its one-hop neighbors. Therefore, in order to make forwarding decisions,
geographic routing protocols, only require the position of the packet’s destination and
the position of its one-hop neighbors. This property has made Geographic Routing
very assuring. Geographic Routing protocols operate according to two functioning
modes: i) the packets are routed from one node to its neighbor, which can be the
closest node to destination or the node that has the highest energy, etc. ii) if there are
no nodes with these prescriptions, the algorithm goes into recovery mode.
Geographic Routing protocols give interesting and attractive solutions [13–18]
for the WSNs because to forward information, a node needs only to know its own
position, and its neighbors and the BS locations. The localization information can be
obtained thanks to the Global Positioning System (GPS). This reduces the complexity
of control management at the level of each node very significantly.
The greedy protocols [18] use the shortest paths to distribute the packets between a
source and a destination node. In the Nearest with Forward Progress (NFP) protocol
[18], the algorithm chooses the nearest neighbor among the nodes that are situ-
ated in the forward direction of the destination to send the packets. In the Most
Forward within the transmission Range (MFR) routing protocol [1], a node selects
the neighbor that is nearest from the destination as the following routing node. In the
case of the normalized advance (NADV) for geographic routing protocol, the neigh-
boring nodes of the source node are selected taking into account their proximity as
well as the cost of the link.
The protocol proposed here is based on EEGR [19] and brings improvements to
it:
– The next neighbor is calculated each round rather than calculating it for each
forwarded packet,
– In order to forward packets around a void through the optimal path, a new
mechanism is used,
– In the proposed approach, the algorithm tries to find a global optimum path without
any assumptions.
1138 K. Redjimi et al.
A WSN is represented as a graph G = (E, V), where E is the set of edges that
corresponds to the links between nodes and V the nodes. The proposed protocol
adopts the model of Ivan Stojmenovic et al. [20, 21] that generalized the energy
model of Rodoplu and Meng’s [22] by adjusting a linear factor for the transmission
and the reception of a radio signal between two nodes separated by a distance d:
u(d) = ad a + c (1)
A node that encounters a void must expand its subgraph by δ degrees to include
other neighbors; this zone is called the recovery area. The same algorithm is applied
by using the source routing to route packets. According to the shortest path that is
included in the packet’s header, the next hop transmits the packet.
In greedy routing techniques, each node participates for packet forwarding towards
the BS. At the network setup time, the nodes select their next forwarding neighbor.
This neighbor forwards all received packets. In order to balance the load of neigh-
boring nodes, the selection is initialized at each round. The duration of a round
depends on the intensity of the network traffic. Note that a node plays the rotation role
if its next forwarding neighbor energy decreases. In each node, a table that contains
the neighbor’s identifier, position and remaining energy is periodically updated. In
addition, each node periodically broadcasts a message displaying its position and
remaining energy.
In the local sub-graph G = (E ,V ), V corresponds to the set of neighbors in
ARouting_Area/ARecovery_Area, E to the set of edges between these neighbors.
The distance between two nodes u and v is identified as dist(u, v), and r is the sensor
node’s communication range.
1140 K. Redjimi et al.
To balance and to regularly distribute the consumption energy on all the neighbors,
we add a new metric. The energy required for the transmission and the reception of
data between two nodes is represented in Eq. (2):
(ad a + c)
Edgeweight = (2)
R Eu
In the WSN, data packets are forwarded from source nodes to the BS via multi-hop
ways. In the proposed approach, we distinguish two modes for packets forwarding:
In the normal mode, the packets are forwarded through the Routing Area whereas
in the recovery mode, the packets are routed through the Recovery Area. A packet
is forwarded in normal mode but when it encounters a void at a node p, it will be
balanced to recovery mode to forward it around the void. When the packet gets to a
node q having at least a neighbor closer to the BS than node p, the packet is returned
back to greedy mode forwarding again. In that case, the packet is forwarded to that
neighbor. To do that, a node p that balances a packet into recovery mode includes in
the packet’s header its position, i.e. the position where the packet was balanced to
recovery mode. The node p includes also in the header the minimum path through
which the packet will be forwarded in recovery mode. Afterwards, each node q that
receives the packet checks first whether it has at least a neighbor closer to the BS than
node p using the position included in the packet’s header. If so, it balances the packet
to normal mode and forwards the packet to that neighbor. Otherwise, it forwards
the packet according to the minimum path included in the header. The algorithm 1
presented bellow summarizes the routing strategy of this approach.
Improved Geographic Routing Protocol for Wireless Sensor Networks 1141
Table 1 Simulation
Parameters Values
parameters
Network nodes number 20 to 160
Generation mode Randomized
Trigger nodes 05 trigger nodes
Size of the network 100 m × 100 m
Transmission range of nodes 30 m
Radio Transmission power 0.028 w
Radio Reception power 0.036 w
MAC sub-layer MAC IEEE 802.11
Angle δ 2π/3
1142 K. Redjimi et al.
Fig. 2 Average of the energy consumed compared to the number of nodes in the network
Improved Geographic Routing Protocol for Wireless Sensor Networks 1143
and the BS is optimal, which is not the fact in the reality. In the other protocols
(GPSR, NFP, and MFR), a neighbor closest to the BS is selected without regarding
the energy consumption criteria. Therefore, the forwarded packets can be expensive
in terms of energy consumption, especially in the case of voids (GPSR protocol
applies recovery mechanism to get out of voids).
5 Conclusion
The proposed approach (called IGR: Improved Greedy Routing geographic protocol)
allows an efficient energy management for the geographic routing protocol. As a solu-
tion to resolve the local minima problem by creating an additive area called recovery
area to forward packets. Mechanisms are introduced to elaborate the optimal way
during each round. After realizing several simulations and analyzing the simula-
tion results, we observe that the proposed protocol contributes to the reduction of
the energy consumption at the sensor nodes and consequently, increase the network
lifetime.
In future works, we project to realize a series of simulations aiming to demonstrate
the effectiveness and the efficiency of this protocol in increasing the packet’s delivery
rate especially in the case of complex networks.
1144 K. Redjimi et al.
References
1. Obaidat MS, Misra S (2014) Principles of wireless sensor networks. Cambridge University
Press, Cambridge
2. Dhobley A, Ghodichor NA, Golait SS (2015) An overview of wireless sensor networks for
health monitoring in hospitals via mobile. Int J Adv Res Comput Commun Eng 4(1):169–171
3. Belghith A, Obaidat MS (2016) Wireless sensor networks applications to smart homes and
cities. In: Smart cities and homes, pp 17–40. https://doi.org/10.1016/b978-0-12-803454-5.000
02-x
4. Lasek A, Cercone N, Saunders J (2016) Smart restaurants: survey on customer demand and
sales forecasting. In: Smart cities and homes, pp 361–386. https://doi.org/10.1016/b978-0-12-
803454-5.00017-1
5. Kandris D, Nakas C, Vomvas D, Koulouras G (2020) Applications of wireless sensor networks:
an up-to-date survey. Appl Syst Innov 3(1):14
6. Khan JA, Qureshi HK, Iqbal A (2015) Energy management in wireless sensor networks: a
survey. Comput Electr Eng 41:159–176. https://doi.org/10.1016/j.compeleceng.2014.06.009
7. Bhushan B, Sahoo G (2018) Routing protocols in wireless sensor networks. Studies Compu-
tational Intelligence, pp 215–248. https://doi.org/10.1007/978-3-662-57277-1_10
8. Boukerche A, Turgut B, Aydin N, Ahmad MZ, Boloni L, Turgut D (2011) Routing protocols
in ad hoc networks: a survey. Comput Netw 55:3032–3080. https://doi.org/10.1016/j.comnet.
2011.05.010
9. Mulert J, Welch I, Seah WKG (2012) Security threats and solutions in manets: a case study
using aodv and saodv. J Netw Comput Appl 35(4):1249–1259. https://doi.org/10.1016/j.jnca.
2012.01.019
10. Ade SA, Tijare PA (2010) Performance comparison of AODV, DSDV, OLSR and DSR routing
protocols in mobile ad hoc networks. Int J Inf Technol Knowl Manag 2(2):545–548
11. Tyagi S, Kumar N (2013) A systematic review on clustering and routing techniques based upon
LEACH protocol for wireless sensor networks. J Netw Comput Appl 36(2):623–645. https://
doi.org/10.1016/j.jnca.2012.12.001
12. Karp B, Kung H (2000) GPSR: greedy perimeter stateless routing for wireless networks. In:
Proceedings of the sixth annual international conference on mobile computing and networking.
ACM Press, pp 243–254. https://doi.org/10.1145/345910.345953
13. Milocco RH, Costantini H, Boumerdassi S (2014) Improved geographic routing in sensor
networks subjected to localization errors. Ad Hoc Netw 13:476–486. https://doi.org/10.1016/
j.adhoc.2013.10.001
14. Lee S, Bhattacharjee B, Banerjee S, Han B (2010) A general framework for efficient geographic
routing in wireless networks. Comput Netw 54(5):844–861. https://doi.org/10.1016/j.comnet.
2009.09.013
15. Boulaiche M, Bouallouche-Medjkoune L (2015) EGGR: energy-aware and delivery guarantee
geographic routing protocol. Wireless Netw 21(6):1765–1774. https://doi.org/10.1007/s11276-
014-0880-1
16. Tao S, Ananda AL, Chan MC (2010) Greedy face routing with face identification support
wireless networks. Comput Netw 54(18):3431–3448. https://doi.org/10.1016/j.comnet.2010.
07.004
17. Kleerekoper A, Filer NP (2015) Perfect link routing for energy efficient forwarding in
geographic routing. Ad Hoc Netw 30:46–62. https://doi.org/10.1016/j.adhoc.2015.03.001
18. Al-Shugran M, Ghazali O, Hassan S, Nisar K, Arif ASM (2013) A qualitative comparison
evaluation of the greedy forwarding strategies in mobile ad hoc network. J Netw Comput Appl
36(2):887–897. https://doi.org/10.1016/j.jnca.2012.10.008
19. Zhang H, Shen H (2007) Eegr: energy-efficient geographic routing in wireless sensor networks.
In: 2007 international conference on parallel processing (ICPP 2007). IEEE, p 67. https://doi.
org/10.1109/ICPP.2007.37
20. Bose P, Morin P, Stojmenović I, Urrutia J (2001) Routing with guaranteed delivery in ad hoc
wireless networks. Wireless Netw 7(6):609–616. https://doi.org/10.1023/A:1012319418150
Improved Geographic Routing Protocol for Wireless Sensor Networks 1145
21. Stojmenovic I, Lin X (2001) Power-aware localized routing in wireless networks. IEEE Trans
Parallel Distrib Syst 12(11):1122–1133. https://doi.org/10.1109/71.969123
22. Rodoplu V, Meng TH (1999) Minimum energy mobile wireless networks. IEEE J Sel Areas
Commun 17(8):1333–1344. https://doi.org/10.1109/49.779917
23. Chen D, Varshney PK (2007) A survey of void handling techniques for geographic routing
in wireless networks. IEEE Commun Surv Tutor 9(1):50–67. https://doi.org/10.1109/COMST.
2007.358971
24. Prim RC (1957) Shortest connection networks and some generalizations. Bell Syst Tech J
36(6):1389–1401
25. https://www.physiome.org/jsim/. Accessed 8 Dec 2020
Supervision and Monitoring
of Photovoltaic Systems Using Siemens
PLC and HMI
Abstract Automation is the top priority in modern industries, the debate on this
phenomenon is always controversial as the machine tends to replace man in several
tasks, and he unfortunately cannot perform the tasks that the machine does in several
fields. The supervision and monitoring of the industrial systems are two very impor-
tant fields in industrial automation. For ensuring the reliability and stability of PV
system operation, it is indispensable to install supervision and monitoring systems.
The work in this paper presents a simple platform for supervision and control in real-
time of pumping system parameters (current, voltage, irradiance, temperature and
flow) based on programmable logic controller (PLC) and human-machine interface
(HMI) of Siemens tools.
1 Introduction
The production technology and the complexity of the operations to be performed, lead
to the implementation of devices and systems for the automation of manufacturing
or production workshops. Automation is the scientific and technological field that
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1147
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_105
1148 A. Bouraiou et al.
performs and controls technical tasks, by machines operating without human inter-
vention, or with the aid of reduced intervention. The supervision and monitoring of
the industrial systems are two very important fields in industrial automation. They
are based on the application of different analysis of many signals and parameters
like temperature, current, vibrations, voltage, etc. Many research works have been
done on this area like [1–10]. On the other hand, the transition to renewable energy
resources utilization has become an indispensable task due to the negative effect of
conventional sources (fossil) in terms of environmental pollution and global warming
[11, 12]. Many scientific papers have been published on this topic like [13–25]. Solar
energy is one of the most popular renewable energies especially in countries that
characterized by huge solar energy potential [26–28]. The supervision and control
of the photovoltaic system is essential key for ensuring the reliability and stability of
PV system operation [5]. The paper presents a simple platform for supervision and
control of pumping system parameters (Current, voltage, irradiance, temperature and
flow) based on a programmable logic controller (PLC) and human-machine interface
(HMI) of Siemens tools. After this introductory section, the paper is organized as:
The first section is ‘The used software’ which divided into 3 parts: Step 7 software,
WinCC Flexible software and Project Creation and Hardware (HW) configuration.
The second section concerns ‘The developed platform’. At the end of this paper the
conclusion is drawn.
STEP7 is the basic software package for configuring and programming of S7 300 and
S7 400 programmable logic controller. It is part of the SIMATIC software industry.
WinCC Flexible is a software compatible with the STEP7 environment, it offers the
configuration of various operator panels and SCADA system, moreover creating the
graphical interface (HMI) and the variables means being able to read the process
values via the PLC, display them so that the operator can supervise, control and
adjust them.
Supervision and Monitoring of Photovoltaic Systems … 1149
Before giving the results of this work, it is more suitable to provide the following
detail: The CPU 314C-2 PN/DP PLC (Siemens) is used for programming using
Ladder language the process step, the main component of this PLC is given Table 1
below:
1150 A. Bouraiou et al.
In this work, the five analog input is used for measurement of the parameters like
current, voltage, irradiance, temperature and flow, the detail of these parameters via
symbolic configuration in Step 7 software is presented in Fig. 3.
Supervision and Monitoring of Photovoltaic Systems … 1151
Fig. 4 S7-PLCSIM
interface
The second interface (system) allows the supervision of the system parameters’
values in real-time like current, voltage, irradiance, temperature and flow, as presented
in Fig. 7. An example of real-time plotting curves is illustrated in Fig. 8.
Supervision and Monitoring of Photovoltaic Systems … 1153
4 Conclusion
The supervision and monitoring of the industrial systems are two very important
fields in industrial automation. In this paper, a simple platform based on Siemens
PLC and HMI was realized in order to supervise and monitor the solar pumping
system parameters. The real realization of this platform requires the sensors signal
converting to universal signals such as 4–20 mA or 0–10 V. This study shows that
can use the PLC and HMI for resolving the supervision and monitoring issues in
photovoltaic systems.
References
1. Atoui I, Meradi H, Boulkroune R, Saidi R, Grid A (2013) Fault detection and diagnosis in
rotating machinery by vibration monitoring using FFT and Wavelet techniques. In: 2013 8th
international workshop on systems, signal processing and their applications (WoSSPA). IEEE,
pp 401–406. https://doi.org/10.1109/WoSSPA.2013.6602399
2. da Silva AM, Povinelli RJ, Demerdash NAO (2008) Induction machine broken bar and stator
short-circuit fault diagnostics based on three-phase stator current envelopes. IEEE Trans Ind
Electron 55:1310–1318. https://doi.org/10.1109/TIE.2007.909060
3. Attoui I, Boutasseta N, Fergani N, Oudjani B, Deliou A (2015) Vibration-based bearing fault
diagnosis by an integrated DWT-FFT approach and an adaptive neuro-fuzzy inference system.
In: 2015 3rd international conference on control, engineering & information technology (CEIT).
IEEE, pp 1–6. https://doi.org/10.1109/CEIT.2015.7233098
4. Attoui I, Boudiaf A, Fergani N, Oudjani B, Boutasseta N, Deliou A (2015) Vibration-based
gearbox fault diagnosis by DWPT and PCA approaches and an adaptive neuro-fuzzy inference
system. In: 2015 16th international conference on sciences and techniques of automatic control
and computer engineering (STA). IEEE, pp 234–239. https://doi.org/10.1109/STA.2015.750
5177
5. Madeti SR, Singh SN (2017) Monitoring system for photovoltaic plants: a review. https://doi.
org/10.1016/j.rser.2016.09.088
6. Attoui I, Oudjani B, Fergani N, Bouakkaz S, Bouraiou A (2020) Novel predictive features
using a wrapper model for rolling bearing fault diagnosis based on vibration signal analysis.
Int J Adv Manuf Technol 106:3409–3435. https://doi.org/10.1007/s00170-019-04729-4
7. Attoui I, Fergani N, Boutasseta N, Oudjani B, Bouakkaz MS, Bouraiou A (2021) Multiclass
support vector machine based bearing fault detection using vibration signal analysis. https://
doi.org/10.1007/978-981-15-6403-1_61
8. Attoui I, Boutasseta N, Fergani N (2021) Novel machinery monitoring strategy based on time-
frequency domain similarity measurement with limited labeled data. IEEE Trans Instrum Meas
70:1–8. https://doi.org/10.1109/TIM.2020.3011874
9. Attoui I (2020) Novel fast and automatic condition monitoring strategy based on small amount
of labeled data. IEEE Trans Syst Man Cybern Syst 1–10 (2020). https://doi.org/10.1109/TSMC.
2020.3018102
10. Boutasseta N, Ramdani M, Mekhilef S (2018) Fault-tolerant power extraction strategy for
photovoltaic energy systems. Sol Energy 169:594–606. https://doi.org/10.1016/J.SOLENER.
2018.05.031
11. Bouraiou A, Necaibia A, Boutasseta N, Mekhilef S, Dabou R, Ziane A, Sahouane N, Attoui I,
Mostefaoui M, Touaba O (2020) Status of renewable energy potential and utilization in Algeria.
J Clean Prod 246:119011. https://doi.org/10.1016/J.JCLEPRO.2019.119011
1156 A. Bouraiou et al.
12. Touaba O, Ait Cheikh MS, Slimani ME-A, Bouraiou A, Ziane A, Necaibia A, Harmim A
(2020) Experimental investigation of solar water heater equipped with a solar collector using
waste oil as absorber and working fluid. Sol Energy 199:630–644. https://doi.org/10.1016/j.sol
ener.2020.02.064
13. Attoui I, Omeiri A (2014) Modeling, control and fault diagnosis of an isolated wind energy
conversion system with a self-excited induction generator subject to electrical faults. Energy
Convers Manag 82:11–26. https://doi.org/10.1016/j.enconman.2014.02.068
14. Attoui I, Omeiri A (2014) Contribution to the fault diagnosis of a doubly fed induction gener-
ator for a closed-loop controlled wind turbine system associated with a two-level energy
storage system. Electr Power Components Syst 42:1727–1742. https://doi.org/10.1080/153
25008.2014.950361
15. Attoui I, Omeiri A (2015) Fault diagnosis of an induction generator in a wind energy conversion
system using signal processing techniques. Electr Power Components Syst 43:2262–2275.
https://doi.org/10.1080/15325008.2015.1082161
16. Bouakkaz MS, Boukadoum A, Boudebbouz O, Bouraiou A, Boutasseta N, Attoui I (2020)
ANN based MPPT algorithm design using real operating climatic condition. In: Proceedings
of the 2020 international conference on mathematics and information technology, ICMIT 2020,
pp 159–163. https://doi.org/10.1109/ICMIT47780.2020.9046972
17. Bouakkaz MS (2020) Fuzzy logic based adaptive step hill climbing MPPT algorithm for PV
energy generation systems. In: 2020 International conference on computing and information
technology (ICCIT-1441), pp 248–252
18. Fergani N, Charef A, Attoui I (2015) Desired closed-loop based self-tuning fractional PID
controller for wind turbine speed control. In: 2015 4th international conference on systems and
control, ICSC 2015. https://doi.org/10.1109/ICoSC.2015.7152771
19. Boutasseta N, Bouakkaz MS, Bouraiou A, Attoui I, Fergani N (2020) Practical implementation
of computational algorithms for efficient power conversion in photovoltaic energy generation
systems
20. Abderrezzaq Z, Mohammed M, Ammar N, Nordine S, Rachid D, Ahmed B (2017) Impact of
dust accumulation on PV panel performance in the Saharan region. In: 2017 18th international
conference on sciences and techniques of automatic control and computer engineering (STA),
pp 471–475. https://doi.org/10.1109/STA.2017.8314896
21. Mostefaoui M, Ziane A, Bouraiou A, Khelifi S (2018) Effect of sand dust accumulation on
photovoltaic performance in the Saharan environment: southern Algeria (Adrar). https://doi.
org/10.1007/s11356-018-3496-7
22. Mostefaoui M, Necaibia A, Ziane A, Dabou R, Rouabhia A, Khelifi S, Bouraiou A, Sahouane
N (2018) Importance cleaning of PV modules for grid-connected PV systems in a desert
environment. In: 2018 4th International Conference on Optimization and Applications, pp
1–6. https://doi.org/10.1109/ICOA.2018.8370518
23. Dabou R, Sahouane N, Necaibia A, Mostefaoui M, Bouchafaa F, Rouabhia A, Ziane A,
Bouraiou A (2017) Impact of partial shading and PV array power on the performance of
grid connected PV station. In: 2017 18th International Conference on Sciences and Techniques
of Automatic Control and Computer Engineering, pp 476–481. https://doi.org/10.1109/STA.
2017.8314901
24. Motahhir S, Chouder A, Hammoumi AEl, Benyoucef AS, Ghzizal AEl, Kichou S, Kara K,
Sanjeevikumar P, Silvestre S (2020) Optimal energy harvesting from a multistrings PV gener-
ator based on artificial bee colony algorithm. IEEE Syst J 1–8 (2020). https://doi.org/10.1109/
jsyst.2020.2997744
25. Motahhir S, El Hammoumi A, El Ghzizal A (2020) The most used MPPT algorithms: review
and the suitable low-cost embedded board for each algorithm. https://doi.org/10.1016/j.jclepro.
2019.118983
26. Blal M, Khelifi S, Dabou R, Sahouane N, Slimani A, Rouabhia A, Ziane A, Neçaibia A,
Bouraiou A, Tidjar B (2020) A prediction models for estimating global solar radiation and
evaluation meteorological effect on solar radiation potential under several weather conditions
at the surface of Adrar environment. Measurement 152:107348. https://doi.org/10.1016/J.MEA
SUREMENT.2019.107348
Supervision and Monitoring of Photovoltaic Systems … 1157
Abstract In this paper a hybrid controller for tracking the maximum power point of
photovoltaic modules is proposed. For an accurate modelling, the double diode model
was adopted, and so as to decrease the current undulation the single-ended primary-
inductor-converter was used. The proposed controller is developed by integrating
sliding mode and Artificial Neural Network, the latter has been designed to delivers
an optimal voltage, which corresponds to the voltage of the maximum power point,
whereas the sliding mode was developed to track the signal of the reference voltage
by computing the duty cycle of the converter. The simulation has been carried out
using Simulink software under different environmental conditions, the results show
that the MPPT controller tracks the reference voltage in 80 ms, and exhibits good
performance at brusque variation of temperature and irradiation, in addition the
accuracy of the proposed method was further justified against previous work.
1 Introduction
The photovoltaic energy is the principal source of renewable energy, which gener-
ates electricity by using photovoltaic modules that convert sun energy into elec-
trical power. In fact solar power is more cost-effective than standard electricity.
However, the power supply output depends on the load demand, which evokes
researchers to propose several algorithms that force the photovoltaic panel to works
at its maximum power. The most commonly employed is perturb and observe [1–3]
due to its simplicity of implementation, but it’s unstable and suffers from oscilla-
tions around the maximum power point. So, to overcome this problem, the non-linear
commands are employed such as sliding mode [4, 5], backstepping [6, 7], and fuzzy
logic [8, 9]. So as to increase the accuracy of the tracking, authors in [10], propose
a hybrid method which integrates perturb and observe (PO) loop in sliding mode
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1159
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_106
1160 K. Chennoufi and M. Ferfra
controller (SMC), the latter was conceived to track reference voltage generated with
PO algorithm. The proposed approach illustrates good results. However, there are
still oscillations around the maximum power point.
So as to surmount the fluctuations issue around the maximum power, the present
paper proposes an MPPT controller which combines Artificial Neural Network
(ANN) and sliding mode controller (SMC) using SEPIC converter. The ANN is
conceived to estimate the optimum voltage that generates the maximum power and
the sliding mode is designed to track the desired voltage by changing the duty cycle
of the converter (Fig. 1).
Unlike the previous works which use the one-diode model for PV modules and a
traditional DC-DC converter, the present work adopts the double diode model for an
accurate result, and a SEPIC converter, because it offers additional advantages than
the traditional DC-DC converter.
This paper is structured as follows. The extraction of the module parameters and
the modelling of the converter are presented in the Sect. 2. The Sect. 3 is devoted to
the MPPT controller design. The Sect. 4 is dedicated to the simulation results and
discussions.
The solar power system considered in this work is a PV panel, coupled to a resistive
load with a SEPIC as illustrated in the Fig. 1. The specifications of the PV panel KD
245 are illustrated in Table 1.
Fig. 1 Proposed
photovoltaic system
Maximum Power Point Tracking Using SEPIC Converter … 1161
Table 1 Polycrystalline PV
Parameter Value
panel specifications
Isc (A) 8.91
Voc (V) 36.9
Im (A) 8.23
Vm (V) 29.8
Ns 60
The single ended primary inductor converter is used since it can reach maximum
power with feeble current’s undulation [11], and it is able to work in a buck or a
boost mode [12]. The mathematical modelling of the converter is carried out based
on these states [13, 14] as illustrated in the Fig. 2.
Figure 2-a shows the equivalent circuit of the SEPIC when the switch is on (u
= 1). In this case the output is supplied from the capacitor C3 , and the amount of
current stored in the inductors raises. The input and the capacitor C2 supply energy to
the inductor L1 and to the inductor L2 respectively. By using capacitor and inductor
equations the representation of state space is as follows:
⎡ ⎤ ⎡ ⎤⎡ ⎤ ⎡V ⎤
dIL1 0 0 0 0 IL1 pv
⎢
dt
dVC2 ⎢ ⎥ ⎢
⎥ ⎢0 0 C 0 ⎥⎢ V ⎥ ⎢ 0
1 ⎥ ⎢ L1
⎥
⎢ ⎥ = ⎢ −1 2 ⎥
⎣
dt
⎦ ⎣0 ⎥ ⎢ c2 ⎥ + ⎢ ⎥
dIL2
dt L2
0 0 ⎦ ⎣ IL2 ⎦ ⎣ 0 ⎦
dV0 −1
dt 0 0 0 RC 3
Vo 0
Figure 2-b shows the equivalent circuit of the SEPIC when the switch is off (u =
0). In this case the inductor L1 and L2 are discharged, which raises the energy stored
in the capacitors. By using capacitor and inductor equations the representation of
state space is as follows:
1162 K. Chennoufi and M. Ferfra
⎡ ⎤ ⎡ ⎤⎡
−1
⎤ ⎡V ⎤
dIL1 0 −10
L1 L1 IL1 pv
⎢
dt
dVC2 ⎢ ⎥ ⎢
⎥ ⎢ C1 0 0 0 ⎥ ⎢ V ⎥ ⎢ 0⎥ ⎢ L1
⎥
⎢ ⎥=⎢ 2 ⎥
⎣
dt
⎦ ⎣ 0 0 0 1 ⎥ ⎢ c2 ⎥ + ⎢ ⎥
L2 ⎦ ⎣ L2 ⎦ ⎣ 0 ⎦
dIL2
dt
I
dV0 1 −1 −1 V
dt C3
0 C3 RC3 o
0
The single diode model is widely used for the modelling of photovoltaic modules due
to its simplicity. However, the accuracy of the model may suffer, as the absolute errors
of current and voltage are relatively high [15]. Thus, in order to have an accurate
modelling the double diode model is considered in this work, which is represented
in Fig. 3.
The output current of a solar cell can be represented by the following equation:
V + Rs I V + Rs I V + Rs I
I = Iph − I01 exp − 1 − I02 exp −1 − (5)
a1 V T a2 VT Rp
Where I: Current of PV module, Iph : Photocurrent, I01 and I02 saturation current
for diode 1 and diode 2 respectively, Rp : Cell parallel resistance, Rs : Cell series
resistance, a1 and a2 ideality factor for diode 1 and diode 2 respectively, V: Voltage
of PV module in voltage, VT : Thermal voltage.
Maximum Power Point Tracking Using SEPIC Converter … 1163
Fig. 3 Double-diode
equivalent circuit
3 Proposed Controller
In this section a sliding mode controller (SMC) combined with Artificial Neural
Network (ANN) is designed as shown in Fig. 1. The ANN provides a voltage reference
for any environmental condition that corresponds to maximum power voltage, while
the SMC tracks this reference by computing the duty cycle of the converter.
The Artificial Neural Network is a data analysis and computational algorithm which
consists of a related group of nodes inspired from biological neuron [17]. The ANN
is the most suitable algorithm for the nonlinear system, its architecture consists of
three layers, the input layer collects the external information, the second layer is the
1164 K. Chennoufi and M. Ferfra
hidden layer, contains hidden neurons that receive information from the input layer
and send them to the output layer [18].
In this paper, the input layer was designed with a database of 158 irradiation and
temperature cases, the temperature has been increased by a step of 5 °C, from 5 °C to
60 °C, and the irradiance was increased with a step of 100 W/m2 for each temperature
degrees, whereas the hidden layer has been conceived with 100 neurons. The output
layer has been conceived with one neuron that corresponds to the maximum power
voltage for every 158 irradiation and temperature cases. The ANN model fitting shows
the outputs for the targets for training and test sets. The R-square value was 0.99999
for all training sets, as shown in Fig. 4, which illustrates an excellent relationship
between the output and target sets.
The SMC is designed in order to tracks the maximum power voltage, Vpv and Vref
are the photovoltaic and the reference voltages respectively.
The tracking error (e) is expressed as follows:
IPV − IL1
V̇pv = (9)
C1
Thus:
IPV − IL1
ė = − Vref (10)
C1
S = e − λė (11)
Ṡ = ė − λë (13)
Where λ > 0.
Second derivative of (10) is given by:
İPV − İL1
ë = − V̈ref (14)
C1
dIPV dVPV
İPV = (16)
dVPV dt
By introducing the Eq. (5) in the Eq. (16) the derivative of photovoltaic current
can be expressed as:
⎡
⎤
V+Rs IPV V+Rs IPV
I01 exp a1 VT
I02 exp a2 VT 1 ⎦
İPV = ⎣− − − V̇PV (17)
a1 VT a2 VT RP
1166 K. Chennoufi and M. Ferfra
By replacing Eq. (17) into (14) and combining (13) with (12) the command u(t)
is obtained:
1
u = 1 − VPV − L1 C1 (AS + Bsign(S)) − λ V̇PV − V̇ref + V̈ref + İPV
VC2 + V0
(18)
The tracking accuracy was evaluated using KD 245 module connected to a 40 load
via SEPIC converter which the parameters are shown in Table 3.
Figure 5 shows the temperature and irradiation cases used in this simulation. In
the first interval [0 s, 23 s], the photovoltaic panel operates under the maximum
irradiation, whereas the temperature has been increased to 45 °C. In the next case
[23 s, 48 s], the irradiation drops suddenly to 580 W/m2 , and the temperature drops
in two successive levels to 25 °C. In the last case [48 s, 60 s], the temperature
is maintained constant at 25 °C, while the irradiation has been increased to reach
1000 W/m2 .
(a) (b)
Irradiance (W/m²)
Temperature (K)
Vref (ANN)
Present method
Method proposed by [10]
(a) (b)
Voltage (V)
Power (W)
Time (S) Time (S)
(a) (b)
Current (A)
Voltage (V)
(a) (b)
Voltage (V)
Current (A)
Time (S) Time (S)
5 Conclusion
This paper presents a sliding mode control combined to the Artificial Neural Network
for Maximum Power Point Tracking of a PV system, which consists of a photovoltaic
module, modelled with double diode model and a SEPIC converter connected to a
resistive load. The assessment of the efficiency of the proposed method has been
carried out by simulating the proposed controller in Matlab/Simulink software under
different weather conditions. In one hand the simulation demonstrates its efficiency
and stability against the previous work, in standard test condition. In the other hand,
the proposed approach exhibits a fast reaction to track the reference voltage with good
performance even when the environmental condition changes. In addition there is no
oscillation in the obtained curves and the algorithm can easily implemented, which
justify the interest behind its use.
References
1. Thakran S, Singh J, Garg R, Mahajan, P (2018) Implementation of P&O algorithm for MPPT in
SPV system. In: International conference on power energy, environment and intelligent control
(PEEIC), Greater Noida, India
2. Ishaque K, Salam Z, Lauss G (2014) The performance of perturb and observe and incremental
conductance maximum power point tracking method under dynamic weather conditions. Appl
Energy 119:228–236
3. Yong Y, Zhaob FP (2011) Adaptive perturb and observe MPPT technique for grid connected
photovoltaic inverters. Procedia Eng 23:468–473
4. Islam S, Liu XP (2011) Robust sliding mode control for robot manipulators. IEEE Trans Ind
Electron 58(6):2444–2453
5. Lachtar S, Bouraiou A, Djaafri O, Maouedj R (2019) Smooth sliding mode-based MPPT algo-
rithm for photovoltaic applications. In: 1st global power, energy and communication conference
(IEEE GPECOM 2019), Cappadocia, Turkey
6. Parsa A, Kalhor A, Atashgah MAA (2016) Backstepping control performance enhance-
ment using close loop identification for quadrotor trajectory tracking. Modares Mech Eng
16(11):224–234
7. Zhang Y, Fidan B, Ioannou P (2003) Backstepping control of linear time-varying systems with
known and unknown parameters. IEEE Trans Autom Control 48(11):1908–1925
Maximum Power Point Tracking Using SEPIC Converter … 1169
8. Belcher R, Fattahi J, Hinzer K, Schriemer H (2020) Fuzzy logic based module-level power
electronics for mitigation of rapid cloud shading in photovoltaic systems. In: IEEE Applied
Power Electronics Conference and Exposition (APEC), Hioki, USA
9. Ziane A, Necaibia A, Mostfaoui M, Bouraiou A, Sahouane N, Dabou R (2018) A fuzzy logic
MPPT for three-phase grid-connected PV inverter. In: Twentieth international middle east
power systems conference (MEPCON), Cairo, Egypt
10. Mostafa MR, Saad NH, El-sattar AA (2020) Tracking the maximum power point of PV array
by sliding mode control method. Ain Shams Eng J 11:119–131. https://doi.org/10.1016/j.asej.
2019.09.003
11. Efendi MZ, Murdianto FD, Setiawan RE (2017) Modeling and simulation of MPPT sepic
converter using modified PSO to overcome partial shading impact on DC microgrid system.
In: International electronics symposium on engineering technology and applications, Surabaya,
Indonesia
12. Duran E, Sidrach-de-Cardona M, Galan J, Andujar JM (2008) Comparative analysis of buck-
boost converters used to obtain I-V characteristic curves of photovoltaic modules. In: 2008
IEEE power electronics specialists conference, Rhodes, Greece
13. Kashyap AR, Ahmadi R, Kimball JW (2013) Input voltage control of sepic for maximum
power point tracking. In: Power and energy conference at Illinois (PECI), pp. 30–35. IEEE
14. Erickson RW (2000) Fundamentals of power electronics, 2nd edn. Springer, Heidelberg, p 910
15. Chennoufi K, Ferfra M (2020) Parameters extraction of photovoltaic modules using a combined
analytical - numerical method. In: International conference on cloud computing technologies
and applications (CloudTech), Marrakech, Morocco
16. Chennoufi K, Ferfra M, Mokhlis M (2021) An accurate modelling of photovoltaic modules
based on two-diode model. Renew Energy 167:294–305
17. Nambiar N, Palackal RS, Greeshma KV, Chitra A (2015) PV fed MLI with ANN based
MPPT. In: 2015 international conference on computation of power, energy, information and
communication (ICCPEIC), Chennai
18. Saadi A, Moussi A (2003) Neural network use in the MPPT of photovoltaic pumping system.
Rev Renew Energies ICPWE 4:39–45
Global Maximum Power Point Tracking
Using Genetic Algorithm Combined
with PSO Tuned PID Controller
Abstract Several techniques have been proposed to track the global maximum
power point (GMPP) in the presence of multiple local peaks introduced when the
PV generator is affected by partial shading. Such methods include scanning methods
and advanced optimization-based strategies. In this paper, a Genetic Algorithm based
MPPT, combined with PID controller tuned using PSO, is proposed to track the MPP.
The GA algorithm is applied to give the reference voltage and measured PV voltage
to the PID controller that its parameters were tuned using PSO. The obtained results
give a good tracking performance of the GMPP under STC conditions and partial
shading scenarios.
1 Introduction
Photovoltaic (PV) solar energy is one of the most important renewable energies for the
following reasons: clean and non-polluting energy, economical and maintainable, the
availability of solar radiation in most countries of the world [1, 2]. The photovoltaic
solar system consists of a group of elements which include: photovoltaic solar panels,
charge controller (including MPPT), storage batteries, and the power inverter (for
AC loads). In general; solar photovoltaic energy in terms of load connection into two
types: isolated from the grid (which is a simplified system consisting of solar PV
that feeds a load), or connected to the grid (which is a complex system that includes
solar PV, local load and grid) [3].
The operation of solar photovoltaic modules is dependent on climatic conditions
lake solar radiation and temperature of the cell, which is also considered to have
a non-linear current-voltage characteristic curve. The output power of PV arrays
depends on the characteristic curves (I-V and P-V) which are nonlinear and change
with environmental conditions [4]. The maximum power is calculated by the product
of current and voltage, this product is known as the maximum power point (MPP)
which is not constant. For this purpose, the maximum power point tracking technique
(MPPT) is used to determine and extract the MPP value [5]. The strategy to approach
MPPT varies from method to method which is classified into two classes, conven-
tional MPPT and artificial intelligence-based approaches [6–8]. Conventional MPPT
technique such as the incremental conductance technique (IC), perturb and observe
(P&O) [9], open-circuit voltage (OCV), and short circuit current [10]. On the other
hand, artificial intelligence techniques are different and developed using various
algorithms such as fuzzy logic (FL) [11], Genetic algorithm (GA) [12], Adaptive
Neuro-Fuzzy Inference System (ANFIS) [13–16], Artificial Neural Network (ANN)
[17] and Particle Swarm Optimization (PSO) [18, 19].
In this paper, the proposed method is based on GA and PSO tuned PID
controller, which combines two techniques of artificial intelligence approaches that
are applied to MPPT to track global MPP in solar PV systems operating in different
climatic conditions. Simulations in Matlab/Simulink are carried out to implement
the proposed method through which results are obtained. In the following section,
the PV energy generation system which combines the PV array and power converter
are presented. The proposed MPPT strategy in this work is illustrated in Sect. 3.
Section 4 presents simulation results and gives discussions. Conclusions are given
in the end.
The PV cell has been the subject of modeling studies resulting in the three known
and most used models depending on the application. When Maximum Power Point
v pv + I pv Rs N SS
I pv = I ph − Id − (1)
R p N SS
A step-up converter is a converter that produces a high voltage at its output terminal
than its input voltage. The boost DC-DC converter a type of switched-mode power
supply consisting of a minimum of two switching devices and one energy storage
element, a capacitor, an inductor, and a filter is generally inserted at the output to
reduce the ripple of VR . The boost converter is illustrated in Fig. 2.
Fig. 3 Photovoltaic system developed with GA-based MPPT and PSO tuned PID controller
(proposed)
The proposed MPPT strategy illustrated in Fig. 3 is applied to find the MPP under
regular and PS weather conditions.
The GA is applied to find the GMPP of the PV system under different operating
conditions. Figure 4 presented the flowchart of the consequent steps of the GA-based
MPPT.
Global Maximum power Point Tracking Using Genetic Algorithm … 1175
Fig. 4 Flowchart of GA
based MPPT
The proposed GA combined with the PSO-PID controller is designed and tested under
MATLAB/Simulink using the Kyocera KC200GT PV panel. In this work, changes
in solar radiation and temperature are applied to verify and assess the robustness
of the proposed controller. Irradiation and temperature patterns that are used in the
consequent simulations are shown in Fig. 7.
As demonstrated in Fig. 8, which PV tracking performance when using the genetic
algorithm-based MPPT algorithm combined with PSO tuned PID controller, the
Global Maximum power Point Tracking Using Genetic Algorithm … 1177
Fig. 8 The PV array’s response to a variable scenario: (a) STC, (b) P1, (c) P2, (d) P3
1178 M. S. Bouakkaz et al.
proposed method tracks the global optimal operating point of the PV system in
every scenario efficiently with less tracking time. It is also characterized by fast
convergence speed, high PV tracking efficiency, and low power loss.
4 Conclusions
In this paper, a global MPPT strategy to improve the efficiency of the PV system
under regular and partial shading environments has been suggested. In this work,
the impacts of various partial shading patterns on the output of the PV system is
also investigated. The proposed strategy is developed to track the optimum operating
point of the PV system. The controller is employed in the voltage control loop
to accurately regulate the optimum operating voltage. The proposed MPPT control
procedure provides excellent performances in the tracking of the generated reference
from the genetic algorithm, which is verified by different simulation results presented
in the paper. The present work may be extended by considering the integration of the
proposed method with a PV inverter to build the solar photovoltaic grid-connected
power generation system.
References
8. Mirza AF, Mansoor M, Ling Q (2020) A novel MPPT technique based on Henry gas solubility
optimization. Energy Convers Manag 225:113409. https://doi.org/10.1016/j.enconman.2020.
113409
9. Motahhir S, Chalh A, El Ghzizal A, Derouich A (2018) Development of a low-cost PV system
using an improved INC algorithm and a PV panel proteus model. J Clean Prod 204:355–365.
https://doi.org/10.1016/j.jclepro.2018.08.246
10. Motahhir S, El Hammoumi A, El Ghzizal A (2018) Photovoltaic system with quantitative
comparative between an improved MPPT and existing INC and P&O methods under fast
varying of solar irradiation. Energy Rep 4:341–350. https://doi.org/10.1016/j.egyr.2018.04.003
11. Bouakkaz MS, Boukadoum A, Boudebbouz O, Attoui I, Boutasseta N, Bouraiou A (2020)
Fuzzy logic based adaptive step hill climbing MPPT algorithm for PV energy generation
systems. In: 2020 International conference on computing and information technology (ICCIT-
1441), pp 1–5. IEEE. https://doi.org/10.1109/ICCIT-144147971.2020.9213737
12. Daraban S, Petreus D, Morel C (2014) A novel MPPT (maximum power point tracking) algo-
rithm based on a modified genetic algorithm specialized on tracking the global maximum power
point in photovoltaic systems affected by partial shading. Energy 74:374–388. https://doi.org/
10.1016/j.energy.2014.07.001
13. Attoui I, Boudiaf A, Fergani N, Oudjani B, Boutasseta N, Deliou A (2015) Vibration-based
gearbox fault diagnosis by DWPT and PCA approaches and an adaptive neuro-fuzzy inference
system. In: 2015 16th international conference on sciences and techniques of automatic control
and computer engineering (STA), pp 234–239. IEEE. https://doi.org/10.1109/STA.2015.750
5177
14. Attoui I, Oudjani B, Boutasseta N, Fergani N, Bouakkaz M-S, Bouraiou A (2020) Novel
predictive features using a wrapper model for rolling bearing fault diagnosis based on vibration
signal analysis. Int J Adv Manuf Technol 106:3409–3435. https://doi.org/10.1007/s00170-019-
04729-4
15. Attoui I, Boutasseta N, Fergani N, Oudjani B, Deliou A (2015) Vibration-based bearing fault
diagnosis by an integrated DWT-FFT approach and an adaptive neuro-fuzzy inference system.
In: 3rd international conference on control, engineering and information technology, CEIT
2015. https://doi.org/10.1109/CEIT.2015.7233098
16. Lasheen M, Abdel-Salam M (2018) Maximum power point tracking using hill climbing and
ANFIS techniques for PV applications: a review and a novel hybrid approach. Energy Convers
Manag 171:1002–1019. https://doi.org/10.1016/j.enconman.2018.06.003
17. Bouakkaz MS, Boukadoum A, Boudebbouz O, Bouraiou A, Boutasseta N, Attoui I (2020)
ANN based MPPT algorithm design using real operating climatic condition. In: 2020 2nd
international conference on mathematics and information technology (ICMIT), pp 159–163.
IEEE. https://doi.org/10.1109/ICMIT47780.2020.9046972
18. Boutasseta N, Ramdani M, Mekhilef S (2018) Fault-tolerant power extraction strategy for
photovoltaic energy systems. Solar Energy 169:594–606. https://doi.org/10.1016/j.solener.
2018.05.031
19. Boutasseta N, Bouakkaz MS, Bouraiou A, Necaibia A, Attoui I, Fergani N (2020) Practical
implementation of computational algorithms for efficient power conversion in photovoltaic
energy generation systems. In: 2020 international conference on computing and information
technology (ICCIT-1441), pp 1–5. IEEE. https://doi.org/10.1109/ICCIT-144147971.2020.921
3761
20. Shams I, Saad M, Kok Soon T (2020) Improved team game optimization algorithm based
solar MPPT with fast convergence speed and fast response to load variations. IEEE Trans Ind
Electron 1. https://doi.org/10.1109/tie.2020.3001798
21. Qais MH, Hasanien HM, Alghuwainem S (2020) Parameters extraction of three-diode photo-
voltaic model using computation and Harris Hawks optimization. Energy 195:117040. https://
doi.org/10.1016/j.energy.2020.117040
22. Al-Dhaifallah M, Nassef AM, Rezk H, Nisar KS (2018) Optimal parameter design of fractional
order control based INC-MPPT for PV system. Solar Energy 159:650–664. https://doi.org/10.
1016/j.solener.2017.11.040
1180 M. S. Bouakkaz et al.
23. Attoui I, Omeiri A (2014) Contribution to the fault diagnosis of a doubly fed induction generator
for a closed-loop controlled wind turbine system associated with a two-level energy storage
system. Electr Power Compon Syst 42:1727–1742. https://doi.org/10.1080/15325008.2014.
950361
24. Attoui I, Omeiri A (2015) Fault diagnosis of an induction generator in a wind energy conversion
system using signal processing techniques. Electr Power Compon Syst 43:2262–2275. https://
doi.org/10.1080/15325008.2015.1082161
25. Bouakkaz MS, Boukadoum A, Boudebbouz O, Fergani N, Boutasseta N, Attoui I, Bouraiou A,
Necaibia A (2020) Dynamic performance evaluation and improvement of PV energy generation
systems using moth flame optimization with combined fractional order PID and sliding mode
controller. Solar Energy 199:411–424. https://doi.org/10.1016/j.solener.2020.02.055
Experimental Validation
of a Neuro-Fuzzy Model of Open Circuit
Voltage and Maximum Power Voltage
Case of Amorphous Technology
1 Introduction
The photovoltaic (PV) emulator is an electronic system able to reproduce the current-
voltage output characteristics of photovoltaic generators. It facilitates laboratory
testing and operational evaluation of a photovoltaic system by applying the same
real-world conditions such as radiation and shading to achieve results similar to
reality [1]. The proper functioning of these emulators depends on the quality of
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1181
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_108
1182 B. Bouachrine et al.
The proposed model is obtained from the semiconductors theory. It represents the
electrical behavior of a PV module by a mathematical model with a one exponential
[5]:
V
I = Isc 1 − C1 exp −1 (1)
C2 Voc
For the computation of the currents Isc, Imp as a function of the illuminance G, and
the temperature T of the PV source we use the system of analytical Eqs. (4). The
system of equations is obtained by smoothing the experimental measurements using
the Levenberg-Marquardt algorithm [6, 7].
⎧
⎨ Isc = Iscn −0.00534 + 0.9768 G + 0.0002441 (T − Tn )
G n (4)
⎩ Imp = Impn −0.005973 + 0.9712 G + 0.000476 (T − Tn )
Gn
With:
Tn : Temperature (°C) under standard conditions.
Gn : Sunlight (W/m2 ) under standard conditions.
The system of analytical Eqs. (5) allows for the calculation of the voltages Vmp, Voc
as a function of the illumination G, and the temperature T of the PV source. This
system of equations is obtained by smoothing the experimental measurements using
the Levenberg-Marquardt algorithm [6].
⎧
⎨ Voc = Vocn + 1.067 Vt log G − 0.3017 (T − Tn )
Gn (5)
⎩ Vmp = Vmpn + 0.5284 Vt log G − 0.222 (T − Tn )
Gn
With:
Vocn : Open circuit voltage in (V) under standard conditions.
Vmpn : Voltage at maximum power in (V) in standard conditions.
V t = N s kT /q: Thermodynamic potential of Ns cells connected in series.
1184 B. Bouachrine et al.
In order to improve the accuracy of the photovoltaic model, we have opted for the
neuro-fuzzy approach for the identification of Voc and Vmp. Neuro-fuzzy systems
that combine fuzzy logic and neural networks have proven their efficiency in a variety
of industrial problems. They are useful for the identification of non-linear, complex,
and difficult to model systems [7, 8].
We have consequently chosen to use adaptive neuro-fuzzy inference systems
(Anfis), which does not require a particular form of modeling. This is a Perceptron-
type neural network with 5 layers, where each layer corresponds to the realization
of one step of a fuzzy inference system (SIF) of the Takagi Sugeno type. A Matlab
code is developed for the generation of the ANFIS model according to the flowchart
in Fig. 1.
After several learning tests, we have come up with the simplest and most efficient
architecture of the ANFIS model illustrated in Fig. 2.
It has two inputs and a polynomial output of order 1, each input is represented by
two fuzzy sets of the Gaussian type. The final output is calculated as the average of
the outputs of each rule weighted by the degrees of activation.
Inputs x and y are expressed as a function of the irradiance G and temperature T
of the PV module according to the system of Eq. (6).
Stopping Criterion:
Max. number of iterations
Tolerance
Learning SIF :
Anfis
Hybrid optimization
x y
w1
A1 N
w2
x A2 N
Voc or Vmp
y B1 w3
N
w4
B N
x= G
+ T − Tn
Gn (6)
y = (T + 273.15) log G
(T +273.15) G n
For the validation of the proposed ANFIS model, we used an experimental database
from Green Energy Park in Bengrire [6]. The database consists of experimental I(V)
electrical characteristics captured with a sampling pitch of 10 min along the day.
In this work, we used a set of measurement samples with 195 I(V) characteristics.
The PV technology studied is of the Amorphous/Microcrystalline type (case of the
SHARP solar module NS-F135G5 [3]). Its technical specifications are described in
Table 1.
To evaluate the accuracy of the estimated parameters and the quality of the
smoothing, we have calculated the coefficient of determination Rd and the root mean
square error RMSE [6]. To simulate the different models, we chose the Matlab 2016a
environment.
In order to estimate the speed of execution of each of the models studied, we
evaluated the average calculation time tCPU, necessary for the calculation of the
I(V) characteristics of the set of measures considered. For the processing carried out
during our investigations, we used a PC-type computer equipped with the Intel(R)
Core (TM) i5-4310U CPU @ 2.00 GHz, 2601 MHz, 2 core(s), 4 logic processor(s),
with a RAM memory capacity of 4 GB.
In Fig. 3 the values predicted by the ANFIS model and the analytical model
(5) are compared with the corresponding experimental values of the Amor-
phous/Microcrystalline technology studied.
Amorphous Technology
Fig. 3. Comparison of simulated and measured voltages for the studied technologies.
This part presents some comparison results of the experimental and simulated I(V)
characteristics obtained in the two cases where the voltages Voc and Vmp are
computed by the Analytical model (5) and by the ANFIS model.
Figure 4 shows an example of application results of the studied PV model for the
studied technology Microcrystalline amorphous, and for the following sunlight G
and temperature T values:
• G (W/m2 ) = [566, 600, 635, 669, 701]
• T (°C) = [36.1, 35.9, 35.8, 36.6, 36.9]
The Fig. 5 illustrate an example of application results of the studied PV model for
the studied technology Microcrystalline amorphous, and for the following irradiation
G and temperature T values:
For the experimental database as a whole, results comparing the performance of the
models used in this study are grouped in Table 3.
Analysis of the results in Table 2 shows that the ANFIS model achieves a very
good trade-off in terms of speed of execution.
Table 3 Summary of
Technology Criteria PV model with PV model with
performance and computing
ANFIS analytic (5)
time tCPU
SHARP Rd 0.9974 0.9407
NS-F135G5 RMSE 0.0472 0.2247
Amorphous t CPU (s) 0.0316 0.0082
Experimental Validation of a Neuro-Fuzzy Model … 1189
5 Conclusion
During our investigations, we look for a model that reproduces the characteristics of
the PV panels as well as possible. To achieve this, we compared two types of models,
namely the neuro-fuzzy ANFIS model and an analytical model. For the calculation of
the Voc and Vmp voltages we successfully used the ANFIS neuro-fuzzy system. The
study results show that the proposed model is more accurate than the analytical model
and allows to better emulate the electrical characteristics of the studied photovoltaic
generator.
Acknowledgement This work is funded and supported by the Moroccan Research Institute for
Solar Energy and New Energies (IRESEN).
References
1. Chen C-C, Chang H-C, Kuo C-C, Lin C-C (2013) Programmable energy source emulator for
photovoltaic panels considering partial shadow effect. Energy 54:174–183
2. Ayop R, Tan CW (2017) A comprehensive review on photovoltaic emulator. Renew Sustain
Energy Rev 80:430–452
3. Ram JP, Manghani H, Pillai DS, Babu TS, Miyatake M, Rajasekar N (2018) Analysis on solar
PV emulators: a review. Renew Sustain Energy Rev 81:149–160
4. Fiche technique de 135W/ 125W NS-F135G5/NS-F125G5 Amorphous Silicon/Microcrystalline
Silicon SHARP MIDDLE EAST FZE (SMEF)
5. Rekioua D, Matagne E (2012) Optimization of photovoltaic power systems, modelization,
simulation and control. Springer, Heidelberg. ISBN 978-1-4471-2348-4
6. Idadoub H, Kourchi M, Ajaamoum M, Yousfi D, Rachdy A (2019) Comparison and experimental
validation of three photovoltaic models of four technology types. Int J Tech Phys Probl Eng
11(41):1–10
7. Shabaan S, El-Sebah MIA, Bekhit P (2018) Maximum power point tracking for photovoltaic
solar pump based on ANFIS tuning system. J Electr Syst Inf Technol 5:11–22
8. Kharb RK, Shimi SL, Chatterji S, Ansari MF (2014) Modeling of solar PV module and maximum
power point tracking using ANFIS. Renew Sustain Energy Rev 33:602–612
Simulation of Natural Ventilation
on Building with Solar Chimney Under
Climatic Conditions of Errachidia
Morocco Zone
Nomenclature
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1191
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_109
1192 H. Aboubakr et al.
P Pressure (Pa)
Pm Pressure in reference state of the fluid (ρ0 , T0 ) (Pa)
β Coefficient of thermal expansion (K− 1 )
K Thermal diffusivity of the fluid (m2 · s−1 )
δ Kronecker symbol
Z Height (m)
1 Introduction
The building sector is the largest energy consumer in the world with a 34% share,
ahead of industry (27%) and transports (28%). It is also responsible for about a
third of greenhouse gas emissions. The electricity consumed in the building sector
represent for 53% of the total electricity consumption in the world [1].
In Morocco the building sector has a share of about 36% of the country’s total
energy consumption, with 29% reserved for the residential sector and the rest for the
tertiary sector, with an increase of 41% in 8 years [2].
To reduce energy consumption and CO2 emissions in the building sector, several
countries are adopting increasingly strict thermal regulations. In France, for example,
by 2020, all new buildings must be positive energy [3]. The Thermal Regulation of
Buildings in Morocco (RTBM) aims essentially to improve the following thermal
performances [2]:
• Reduce the heating and air conditioning needs of buildings;
• Improve the comfort of non-air-conditioned buildings;
• Reduce the power of heating and air conditioning equipment to be installed;
• Encouraging architects, engineers and supervisors to use efficient thermal design
approaches to the building envelope;
• Provide builders, public decision-makers and funders with a tool to improve the
productivity of their investments;
• Assist in the realization of energy diagnoses of existing buildings.
To improve thermal comfort in buildings and reduce environmental impact, it’s
necessary to use passive solutions. So integrating natural ventilation into the building
can be an efficient way to reverse the national energy trend in the building.
According to ADEREE, the zoning map of Morocco includes six climatic zones
that are represented by the following cities:
Zone 1: Agadir;
Zone 2: Tangier;
Zone 3: Fez;
Zone 4: Ifrane;
Zone 5: Marrakech;
Zone 6: Errachidia.
Simulation of Natural Ventilation … 1193
This work presents a simulation study of the effect of the natural ventilation of a
room using a solar chimney with different widths under the climatic conditions of the
climatic zone of Errachidia Morocco. The overall goal is to simulate the air velocity
and turbulence of kinetic energy in each room position relative to each width of the
solar chimney, in order to determine the optimum width of the solar chimney to be
used in this climatic zone.
In a study presented by Sajjad et al. [4] the authors have simulated by ANSYS
software the effect of the solar chimney in the natural ventilation of buildings. They
examined the effect of the inclination of the chimney with different angles (30°,
45°, 60°, 90°) and different widths (0.2 m, 0.4 m, 0.6 m, 0.8 m) and the height of the
chimney. The results of this study show that the optimum chimney width to maximize
natural ventilation should be 0.4 m and the length should be greater than 2 m or equal
to half of the length of the room, this for different inclinations except for 90°, Thus
the width of the window must be more than 1 m.
By their study on the thermo-aeraulic phenomena in a solar chimney, Saifi et al.
[5] showed that the temperature gradient between the absorber and the glass varies
according to the incident solar flux. The design adopted has resulted in relatively
high air flows at the outlet, which are favorable for use in natural ventilation. The
variation in the thickness of the air gap (chimney width) plays a very important role
and significantly increases the volume flow rate of the air. For optimal thermal draft,
the angle of inclination must be 45°.
Kumar et al. [6] carried out a comparative study between a vertical and horizontal
solar chimney, their results show that vertical chimney enhanced ventilation rate
much better in comparison to Horizontal chimney as much as 275% enhancement,
and vertical Solar Chimney enhanced air flow stream velocity up to 22 times.
Sudheer et al. [7] studied the dynamic and thermal behaviour inside a solar
chimney for three different chimney widths (0.1 m, 0.2 m, 0.3 m). The study showed
that for a width of 0.1 m, the air velocity and temperature inside the chimney is
higher than the other two cases. Thus in case 1 (0.3 m), some amount of stagnant
air is observed at the air inlet near the absorber. By decreasing the air gap from
0.3 m to 0.1 m nearly 44.18% of outlet velocity increase has been observed. So they
concluded that when the chimney width increases there is a decrease in air velocity.
In their study on passive cooling by a solar chimney in southern Algeria,
Belfuguais et al. [8] determined that the solar ray is the most important param-
eter influencing solar chimney performance. Thus an optimal design between the
chimney width and the absorber must be taken in consideration to increase the air
flow rate in the room.
Charvat et al. [9] simultaneously installed two solar chimneys: one with an
absorber to store heat, the other without absorber, to compare their performances. The
study showed that the air velocity inside the solar chimney containing the absorber
1194 H. Aboubakr et al.
is 25% higher than that of the chimney without absorber this is because the absorber
gives up its heat to overheat the air in order to escape more quickly to the outside.
Kasim et al. [10] investigated the effect of different opening positions on natural
ventilation. The study is based on nine different positions of two openings located on
two opposite walls, the result of this study shows that the air flow rate is maximum
when both openings are located at the top of the air flow level.
Mekkawi et al. [11] simulated natural ventilation through a solar chimney in
Egypt. Their study showed that the installation of the solar chimney of size of (0.4
× 1.5 m) can decrease the temperature from 29.45 °C to 28.64 °C in the room and
increase the air velocity with 50%.
Long Shi [12] carried out a study of the impacts of wind on solar chimney perfor-
mance in a building, his study shows that a higher wind velocity does not represent
a better solar chimney performance, and the effects from the window area keep
decreasing under a higher wind velocity.
Based on the results of all these studies, it can be concluded that there are several
parameters that can affect the air velocity at the outlet of the solar chimney, such as
the intensity of the solar radiation, the width of the chimney, the type of glass used
in the chimney, the type of absorber used, the entrance surface of the chimney, the
high of the chimney and the inclination of the solar chimney.
3 Methodology
3.1 Principle
density difference we use a chimney with glass faces. The glass allows transmitting
the solar rays to the air that passes from the chimney, the air heats up and rejects to the
outside, the air of the room replaces the air of the chimney and the cycle continues.
With this method the room will refresh in a natural and continuous way and without
any energy consumption or noise, as opposed to mechanical cooling systems.
∂ ui
=0 (1)
∂ xi
With:
• Pm = p + ρ0 gz: the pressure corresponding to the reference state of the fluid (ρ0 ,
T0 ), ν the kinematic viscosity and g the gravitational field.
β: the coefficient of thermal expansion at constant pressure. It describes the change
of volume at constant pressure as a function of temperature.
Energy equation:
This equation is deduced from the first principle of thermodynamics by neglecting
the effect of pressure in front of temperature variation. It is then expressed by:
Simulation of Natural Ventilation … 1197
∂(T ) ∂ u j T ∂ ∂T
+ = k (3)
∂t ∂x j ∂x j ∂x j
Where:
k = ρCλ is the thermal diffusivity of the fluid (m2 ·s−1 ), Cp is the heat capacity
p
Air velocities and turbulence of kinetic energy in different plane of the room for the
0.3 m width of the solar chimney (Figs. 4, 5, 6 and 7).
• Air velocities and turbulence of kinetic energy in different plane of the room for
the 0.4 m width of the solar chimney.
• Air velocities and turbulence of kinetic energy in different plane of the room for
the 0.5 m width of the solar chimney
• The results of the study show that the air velocity in the room is higher in the
chimney of 0.5 m of width (Fig. 14), 2,74 m/s in the middle of the plane taken
at 0.2 m (Fig. 13), and this velocity decreases as we move up and down the level
of the window plane. The lowest velocity is for the chimney of 0.4 m in width,
(Fig. 10), 2.12 m/s in the middle of the plane taken at 0.2 m, (Fig. 9). For the
distribution of the air velocity in different zones of the room, it is noted that it is
more important in the case of the chimney of 0.5 m of width (Fig. 14). Thus the
ventilation is present in all zones of the room for the three cases studied (Figs. 6,
10 and 14), this clearly shows the importance of the solar chimney in the natural
ventilation. The results also show that the air velocity in the room is not uniform,
it is higher in the middle of the room for the two chimneys with a width of 0.4 m
(Fig. 9) and 0.5 m (Fig. 14), while for the chimney of 0.3 m of width the air
velocity is higher at the solar chimney (Fig. 4), this is because the air heats up
Simulation of Natural Ventilation … 1199
faster and flows out into this chimney. For the turbulence of the kinetic energy it
is higher in the plane taken at 2 m according to the variations of the wind speed in
the room for the three chimneys, (Figs. 7, 11 and 15); and it decreases by going
up and down of the window plane. In all three cases, the absence of a reverse
flow of air in the solar chimney was reported; this refers to the phenomenon of
the difference in density between inside and outside air (Figs. 4, 8 and 12).
1200 H. Aboubakr et al.
5 Conclusion
References
Abstract As the world’s population increases, the demand for vegetarian products
increases, therefore, in order to preserve the food products throughout the year, it
was necessary to find a way that can preserve almost all the nutritional and vitamin
value. In this term, we find that the most efficient method is indirect solar drying.
This latter is used for drying, because of its many advantages, such as being covered
from direct sunlight, which protects the enzymes in the product, while preserving the
original color and reduce the drying time. However, with all these advantages, this
method had problems with air uniformity along the trays. Therefore, in this work,
we will study the dimensional parameters of the system in order to improve the air
and temperature distribution inside the drying chamber, and to be more precise the
study is mainly based on the optimization of the air inlet which is the air gap of the
solar air heater for a newly designed drying chamber. The ANSYS software is used to
simulate the system, which usually consists of a drying chamber and a solar collector
under the meteorological data of Oujda city. As a result, we have found that the air
distribution in the drying chamber depends directly on the size of the air gap of the
solar air heater.
1 Introduction
Solar energy is one of the most important clean and inexhaustible energies on earth,
as it represents an energy that can be exploited without limit as long as the sun is
present. This energy can be converted into different forms that can be used in daily
life such as electricity, heating, and cooling. Since there is an enormous increase in
the world’s population it becomes necessary to satisfy the need for energy and food,
therefore solar energy has been exploited in other systems related to the conservation
and storage of the alimentary product, such as solar drying, Which is an ancient and
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1205
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_110
1206 M. Salhi et al.
Air outlet
The door of
Solar radiation
the chamber
Air inlet
2 Methodology
The indirect solar dryer studied shown in Fig. 1 composed of a 1.5 m2 solar air heater
that composed of an aluminum absorber painted in black and a simple glazing, the
chamber is actually composed of two parts the first one contains three trays where the
products are distributed and the second one is equipped with solar air heater to make
the air pass horizontally in the trays, so the air distribution will be more uniform [5].
The governing equations are based on the Navier-Stokes equations, while the
turbulent pattern inside the solar dryer is modeled using the K-epsilon model,
which characterized with two equations, the turbulence kinetic energy k (1) and
the dissipation rate of turbulent kinetic ε (2).
∂k ∂(ku) ∂(kv) μt ∂ 2 k ∂ 2k
ρ( + + ) = (μ + )( 2 + 2 ) + G k − ρε (1)
∂t ∂x ∂y σk ∂ x ∂y
∂ε ∂(εu) ∂(εv) μt ∂ 2 ε ∂ 2ε ε
ρ( + + ) = (μ + )( 2 + 2 ) + (c1 G k − c2 ρε) (2)
∂t ∂x ∂y σε ∂ x ∂y k
The simulation was made under the metrological conditions of Oujda city, the data
used in the simulation where collected from the metrological station installed in
university Mohammed first in Oujda presented in figure below. The climate data of
the chosen day which are the ambient temperature and the solar radiation where
transformed into equations as presented in Eqs. (3) and (4) before inserting it in
ANSYS FLUENT as a UDF subroutine (Fig. 2).
1208 M. Salhi et al.
π
Tamb (t) = 25 + 6 cos( (t − 14)) (3)
12
t −6
G sun (t) = 965 sin( π ), 6 < t < 20 (4)
14
The studied solar dryer is working under naturel convection and for that the
bossinessq approximation is used, and the inlet as well the outlet where pressure
type, regarding the walls we have convective heat losses in all walls including the
solar air heater while in the glass cover we have a mixed mode of convective and
radiation losses.
The trays are modeled as a porous medium with 50% porosity, in order to model
the airflow resistance created by the trays.
In order to make sure that the simulations made by ANSYS FLUENT is right, two
validations with experimental and numerical results where made the first article
[6] studied a cabinet dryer equipped with a chimney and a solar air heater, the
outlet temperature of this latter was investigated for different lengths, the comparison
between our results and the authors is presented in Fig. 3.a, the other validation
concerns a wooden chamber equipped with a solar chimney [7] the air velocity at
the outlet is presented in Fig. 3.b in those two validations our results showed a better
values than the numerical ones found by the same authors.
Investigating the Effect of the Air Gap of a Solar Air Heater … 1209
Fig. 3 Comparison between our results and the reference [] (a) and [] (b)
3 Results
Fig. 4 The mean air velocity and temperature at the inlet of the chamber for different inlet size at
1:00 pm
Fig. 5. a The average air temperature in the drying chamber, and b the mass flow rate at the inlet
of the chamber in function of the air gap size at 1:00 pm.
To see more clearly the air velocity and temperature along the three trays, we
have presented in Fig. 6 the average air velocity and temperature in the trays for
different air gap sizes, and the results show that the air flow distribution along the
trays is better for an inlet size equal to 7 cm. although for this size, the velocity and
temperature at the inlet is low compared to 4 cm. This is due to the amount of the
airflow entering the drying chamber which becomes high for the inlet size 7 cm.
From Figs. 4, 5 and 6, it can be concluded that the increase of the inlet size
provokes an increase in the air mass flowrate as a results the efficiency of the solar
drying system.
Investigating the Effect of the Air Gap of a Solar Air Heater … 1211
Fig. 6. The mean air velocity (a) and temperature (b) in the three trays for an inlet size from 3 to
7 cm at 1:00 pm.
4 Conclusion
References
6. Ghaffari A, Mehdi R (2015) Modeling and improving the performance of cabinet solar dryer
using computational fluid dynamics. Int J Food Eng 11(2):157–172
7. Jyotirmay M, Sanjay MA (2006) Summer performance of inclined roof solar chimney for natural
ventilation. Energy Build 38:1156–1163
Experimental Investigation of Efficiency
and Dynamic Losses for a Constructed
Solar Boost Converter
1 Introduction
Up to now, power converters have been ushering in a new kind of industrial revolution.
They have been extensively indispensable in manifold industrial applications, notably
for the progress of renewable energy resources [1, 2].
Therefore, the DC-DC converter has been spotlighted in multiples studies for
diverse intentions. It provides good operability, simple and affordable design
to implement. However, the widespread shortcoming of these converters is the
switching losses that arise in the hard switching operating mode [3]. Above and
beyond, higher switching frequency gets the higher switching losses along with elec-
tromagnetic interface. Yet, renewable energy applications are one of the principal
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1213
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_111
1214 W. El Merrassi et al.
applications that rely on the converter, however these class of resources known by
their significantly lower yield. Accordingly, this type of applications involve specific
design criteria to be adopted such as diminished EMI, minimized power losses,
and stabilized junction temperature. In order to break this bottleneck, and enhance
the efficiency of the converter, soft switching strategies are incited as an alterna-
tive to hard-switching operating mode. Several snubbers have indorsed for different
converter to eschew the high losses, and limited the rate change in voltage/current
and overvoltage during the turn on/off. The snubbers presented in the literature can be
classified into active and passive snubbers. As presented in [4, 5] the active snubber
are used to reduce the turn-on switching losses. The auxiliary circuit proposed for
the converters use an extra switch that includes an intermediate control circuit. Some
of the problems that persist for the active snubber is the synchronicity issue of the
control for the different switches involved in the converter, which rise the complexity
of the control procedure, the extend size of the circuit, and the cost [6]. There-
fore, the passive snubber have been a worthwhile solution that offer a low cost,
easiness of control and implementation. Yet, the passive snubber presented in the
literature has unveiled some downsides highlighted under the inability to promise a
perfect turn-off zero voltage switching, regardless of the switching turn-on at zero
current [7]. Accordingly in [2, 6, 8], the both models was not able to eradicate
the switching losses. In addition, in [9–11] the increase number of passif compo-
nents in snubber cells engender extra conduction losses, intricated applications, and
expensive. Furthermore, certain passive snubbers are not qualified for the family of
fundamental DC-DC converter, they are notably provided for their own converter
structure [12]. Consequently, the original idea for proposing new snubbers and high-
performant solar converters primarily originates from fixing the shortcomings and
improving the efficiency of previous approaches snubbers and converters. Conse-
quently, the snubbers circuits and soft-switching converter ought to achieve an equi-
librium amid the reliability and easiness of integration. In this treatise, the dynamical
losses for a conventional solar converter is explored, to promote the system’s perfor-
mance by lessening the total losses of the solar DC-DC boost converter. A novel
approach to design RID and RCD snubber cells had been proposed. The snubber
assured a zero voltage-switching in the switch-on periods side to a zero current-
switching in switch-off periods, which affords a perfect soft-switching for the main
interruptor in the both transition periods. The proposed snubber cells entail merely
few passive components; it has a simple structure with a minor size, easiness of drive
and low cost. The structure of this paper presented as follow. In the first hand, the
fundamental operational principles of the suggested snubber are proved in depth,
furthermore the experimental test is provided to verify the results’s efficiency above
to a comparison of the effictiveness of the system face to other proposed models in
literature.
Experimental Investigation of Efficiency and Dynamic Losses ... 1215
The power losses in nearly all converters are leading out of two major bases: Conduc-
tion losses and dynamic losses. The conduction losses in a boost converter take place
when the switch is entirely in conduction. The resistance elements persisted in the
different component of the boost converter such as diode, inductors, capacitors, and
switch are the origin of these losses. The conduction losses is outlined as [13]:
Where, Rgen is the generator resistor, Rgsl is the induction resistor, Ron is the switch
resistor, Rd is the diode resistor, Resrc is the capacitor resitor, IL.eff is the effective
current that cross the generator and inductance, I.Tr.eff is the effective current that cross
the main switch, I.d.eff is the effective current that cross the diode, I.d.av is the average
current of the diode, I.c.eff is the effective current that cross the output capacitor of
the converter.
Dynamic losses in the boost converter include three sources of losses represented
as: choke core, Switch device, and diode. However, the losses in Switch depends
majorlyon losses in gate, output capacitance and switch losses during transition
period. The mathematical expression of these losses are introduced below:
1
Psc = Cs V 2 f sw (3)
2 out ds
Where, Psg , Psc are the losses in the switch, Psw is losses in switching periods,
Pdiode is the losses produced by the diode, Pcor is the losses produced by the choke
core.
1216 W. El Merrassi et al.
The total dynamic losses are equavilent to the summation of individual switching
losses of converter elements and they can be described as follows:
In spite of everything, the power losses in the core choke, diode, gate control and
the output capacitance can be marginal against the losses occur the switching periods.
Which link directly the dynamic losses in the converter to the switching losses in
the switch. Therefore,in the boost converter, the Semiconductors represent the most
element presentinglosses in the circuit. Theswitchon/off periodsoftheswitch is in the
order of a few tens of nanoseconds to multifold microseconds. At these switching
transitions moment, very large power losses can occur in the semiconductor device.
Despite, the switching time of semiconductor elements is very brief, the average
power loss can be substantive. Dynamic losses are less depending on power load but
instantly depending on switching frequency as shown in the Eq. (2)–(4). As Fig. 1
unveils, the dynamic losses in the switch are predominant compared to the conduction
losses. For this reason, reducing the switch losses will gain less power loss by the
circuit.
The suggested passive snubber is served to conserve the main switch’s high perfor-
mance of the converter as shown in Fig. 2. The snubber contains two cells, the first
one RID is composed with a diode and resistor with a parallel inductor implemented
in series to the switch, while the second RCD cell comprises a parallel diode and
resistor in series with a capacitor [13].
The capacitor C and the inductor L delay the time at which the switch reaches
the final value of voltage and current. In return, the value of the capacitor and the
inductor usually optimized in order to minimize the power dissipated in the switch.
However, it may not be the best value of the snubber elements from the reliability
Experimental Investigation of Efficiency and Dynamic Losses ... 1217
point of view. The stress in switch reduced but the losses persist. Therefore, in the
proposed snubber circuit, the elements are selected based on maximum reliability
considerations for the set of the switch and its snubber. To ensure as results a soft-
switching and a minimum power loss for the switch. As shown in Fig. 2b, in this
circuit when the interruptor proceeds turning-off, load current charges the capacitor
C up to source voltage V. Consequently, it delays the time at which the capacitor
(alsoswitch) voltage waveform reaches the final value V, the snubber capacitor allows
the voltage to rise and reach full value V after the switch current falls to zero. Besides,
the same scenario achieved by the snubber inductor that allows the current to rise
and reach full value I after the switch voltage falls to zero.
The performance characteristics of both the conventional and the proposed boost
converter are verified by the prototype of 1.6 kw and 20 Khz switching operation
using the same components and the same conditions [13]. The prototype itself is
summarized, with the aforementioned design considerations. The converter operated
with a source simulating PV as a source.
The signal gate is given to the devices through an intermediate interference circuit
that provides isolation from the hardware power circuit and the control circuit, as
shown in Fig. 3(b). This circuit integrated a 6N137 Optocoupler to isolate, and an
IR2112 Driver, which had as role amplifying the output signals, delivered from the
control card. From a low voltage to reach the minimum gate threshold voltage of
1218 W. El Merrassi et al.
Fig. 3 Galvanic-isolation circuit a Hardware set-up of the boost converter with the galvanic-
isolation circuit. b the switch signal in the output of the 6N137 Optocoupler and IR2112 driver
the MOSFET. The Fig. 3(b) Channel 2 Shows the digital output waveform of gate
pulse for the boost converter Mosfet switch generated by the DSP, which offers a
signal variation of 0–3.3 V, the switching frequency is 20 kHz equals to the imposed
frequency 20 kHz and the duty cycle is about 50%. The output signal of Optocoupler
is amplified from 3 V to 12 V and the others parameters of switching frequency and
duty cycle remain the same as before as it can be seen in channel 1 in the Fig. 3(b)
that illustrates the output of IR2112 Driver for boost converter [13].
In the proposed implementation, we tested our prototype using a low power supply.
Prior to experiments, the Fig. 4, 5 and 6 present the experimental results of the
active switch’s performance for a solar boost converter operated in both mode hard-
switching and soft-switching conditions. As illustrated in Fig. 4(a)–(c) the drain-
source voltage in the active switch for the conventional circuit endows a high voltage
and current values along with high spike during the device’s turn-off. And, that
is linked to the fast changing of the voltage dv/dt. The proposed circuit achieved
a soft switching with a zero voltage switch at turn-off. Due to the control of Rs
Cs that improved theperformance significantly by reducing the changingrate dv/dt,
which raised the switch’s voltage gradually from0 toV0, as shows in Fig. 4(b)–(d).
The overlapped area between the voltage and current switch is almost omitted. As
results, the main switch in the proposed circuit had zero switching point at the turn off
and turn on. Consequently, it exhibits zero power losses during turn-on and turn-off.
On this ground, the switching loss of the advanced circuit is reduced to a lower loss
compared to the one in the conventional boost converter with hard switching operating
mode, owing to the deduction of the dv/dt and di/dt. The lessen losses reach 65.55%
Experimental Investigation of Efficiency and Dynamic Losses ... 1219
Fig. 4 The voltage and current waveforms for the active switch in a, c the hard switching operating
mode, b, d the soft switching operating mode
27.0
60
i
i
Temperature (°C)
i 50 i 24.0
40
21.0
i
i 30
i 18.0
i 20
i 10 i 15.0
i 83 i 166 i 245 i 332 i 415 i 498 i 100 i 200 i 300 i 400 i 500
Time (s) (a) Time (s)
(b)
i
i
Fig. 5 The Switch junction temperatures measured for a the conventional boost and, b the proposed
boost
in the switch. From the Fig. 5(a), it can be observed that hard-switching operating
mode increases the temperature in the Switch up to 58 °C due to the presence of
the spikes, ripples as was seen in Fig. 4(c) and the high-power losses, which impact
the switch’s performance and accelerates aging of the device. Comparing to that, the
proposed circuit has unveiled a advanced dynamic behavior of the temperature by
1220 W. El Merrassi et al.
reducing it into 44.44% as given in Fig. 5(b). Owing to the fewer power losses and
the good waveforms of the voltage and current above the switch.
To analyze further the effectiveness of the proposed converter, and examine this
issue with the previous proposed models, an electrical characteristics analysis of the
principal components is adopted for the different converters. As seen in Table 1, the
major drawback of zero voltage switching of boost converter in [9] is the persistent
of the ripple and peak in the main switch current. This issue would raises the turn-off
capacitance losses, which enlarge the global losses in the system. Further, to the need
of an additional control circuit for the snubber circuit that expand the size and the cost
of the converter. Alongside in [6, 11] the zero voltage transition converter includes a
high number of extra component, which diminished the efficiency, complicated the
implementation and increased the cost. In addition, the high voltage rate of power
main switch enlarge the on-state resistance Rds(on). As result, in the converter [14]
the conduction losses of the main switch get increased permanently. However In
[15], the current and voltage stress of the main switch is substantially decreased.
On the ground of this, in this converter, the switch-on capacitive losses are atomic.
Compared totheseconverters,theproposed model provided a good voltage and current
quality waveform, by assuring less of stresson the switch, higher power efficiency,
ease of control and implementation and a very low cost.
Table 1 Comparativeresultsforthedifferentconverter
Converter Converter Converter Converter Converter Converter Proposed
[9] [6] [14] [15] [11] [10] converter
Number of 6 9 5 5 7 5 5
extra
component
Type of Active Passive Passive Passive Passive Active Passive
snubber
Voltage stress VH 1.6VH 1.2 VH VH VH 1.3 VH VH
of the main
switch
The main 0 1.5Vds 0.2 Vds 0 0 5.5 Vds 0
switch peak
voltage (V)
The main 2 Ids,rms 0 5 Ids,rms 0 1.2 Ids,rms 1.8 Ids,rms 0
switch peak
current (A)
Additional Yes No No No No Yes No
control circuit
Efficiency (%) 95% 92.8% 95.6% 94.2% 94% 93.3% 97.3%
Cost High High Medium Medium High Medium Low
Implementation Most Most Simple Complex Complex Simple Simplest
complexity complex complex
Experimental Investigation of Efficiency and Dynamic Losses ... 1221
97
Ref[15]
Efficiency (%)
92
Proposed
87
Ref[14]
82 Ref[11]
77 Ref[9]
20 40 60 80 100
Conventional
Output Power (%)
Fig. 6 Efficiency curves of the proposed converter under hard switching and soft switching,
compared with the soft converter proposed by [9, 11, 14, 15]
The efficiency curves of the conventional solar hard-switching boost converter and
the proposed soft boost converter are illustrated in Fig. 6. In total, the soft switching
converter presented an extremely high power efficiency than that of the hard switching
converter for the different power range. Overall, the value of the efficiency attained is
97.03%, while the conventional converter reached 90.17%. Therefore, the proposed
circuit enhanced, and increased the efficiency up to 7%. On another hand, the figure
presented as well the power efficiency for the previous model converter presented in
the literature [9, 11]. These models presented different value of the power efficiency.
For the [6] model the efficiency gained for full load is 95.3%, while in [9, 11, 15] it
ranges among 92% and 94.2%. Compared to these results, the proposed converter still
achieves higher efficiency, the improvement reached over 6.83% for light load and
more than 4.51% for heavy load. As advantage, the converter assured high efficiency
for light loads as for heavy one.
5 Conclusion
References
1. Solatialkaran D, Zare F, Saha TK, Sharma R (2020) A Novel approach in filter design for grid-
connected inverters used in renewable energy systems. IEEE Trans Sustain Energy 11:154–164
2. Shan Y, Hu J, Chan KW, Fu Q, Guerrero JM (2019) Model predictive control of bidirectional
DC–DC converters and AC/DC interlinking converters—a new control method for PV-wind-
battery microgrids. IEEE Trans Sustain Energy 10:1823–1833
3. EL Merrassi W, Abounada A, Ramzi M (2019) Design of a PWM sliding mode voltage
controller of a DC-DC boost converter in CCM at variable conditions. In: Advances in smart
technologies applications and case studies. Springer, Cham, pp 263–270
4. Jabbari M, Mokhtari M (2020) High-Frequency Resonant ZVS boost converter with grounded
switches and continuous input current. IEEE Trans Industr Electron 67:1059–1067
5. Mohammadi MR, Peyman H, Yazdani MR, Mirtalaei SMM (2020) A ZVT bidirectional
converter with coupled-filter-inductor and elimination of input current notches. IEEE Trans
Industr Electron 67:7461–7469
6. Zhang Z, Yang X, Zheng TQ, Zhang J (2020) A Passive soft-switching snubber with energy
active recovery circuit for PWM inverters. IEEE Access 8:100031–100043
7. Chen Y-T, Li Z-M, Liang R-H (2018) A Novel Soft-Switching interleaved coupled-inductor
boost converter with only single auxiliary circuit. IEEE Trans Power Electron 33:2267–2281
8. Lin H, Hu B, Li F, Chen J, Si L, Zhou X, Li Y, Chen J, Yan M, Dong Y (2018) A fault-tolerant
two-permanent magnet synchronous motor drive with field-oriented control scheme. In: 2018
2nd IEEE advanced information management,communicates,electronic and automation control
conference (IMCEC), pp 1029–1033
9. Zhang Y, Cheng X-F, Yin C, Cheng S (2018) Analysis and research of a soft-switching
bidirectional DC–DC converter without auxiliary switches. IEEE Trans Industr Electron
65:1196–1204
10. Muchina EG, Masike L, Njoroge Gitau M (2019) High boost-ratio bidirectional converter
for interfacing low-voltage battery energy storage system to a DC bus. IET Power Electron
12:2372–2381
11. Do H-L (2011) Nonisolated bidirectional zero-voltage-switching DC–DC converter. IEEE
Trans Power Electron 26:2563–2569
12. Meng T, Ben H, Li C (2019) An input-series flyback converter with the coupled-inductor-based
passive snubber for high-input voltage multiple-output applications. IEEE Trans Ind Electron
66:4344–4355
13. El Merrassi W, Abounada A, Ramzi M (2020) Switching losses analysis of a constructed solar
DC-DC static boost converter. Adv Electr Electron Eng 18:132–141
14. Lin J-Y, Lee S-Y, Ting C-Y, Syu F-C (2019) Active-clamp forward converter with lossless-
snubber on secondary-side. IEEE Trans Power Electron 34:7650–7661
15. Zhang Y, Cheng X-F, Yin C, Cheng S (2018) A soft-switching bidirectional DC–DC converter
for the battery super-capacitor hybrid energy storage system. IEEE Trans Industr Electron
65:7856–7865
16. El Merrassi W, Abounada A, Ramzi M (2021) Advanced speed sensorless control strategy for
induction machine based on neuro-MRAS observer. Mater Today: Proc
Improvement of the Non-linear Control
Strategy of a Wind Turbine
by a High-Gain Observer
Abstract The production of a wind turbine depends mainly on the energy of the
wind, the power graph of the turbine, and the response of the generator to changes
in the wind. This article suggests a robust sliding mode of second-order nonlinear
control technique based on the concept of the super-twisting algorithm for the doubly-
fed induction generator (DFIG) of a wind turbine that provides optimal power produc-
tion (efficiency). For this task, an observer with a high-gain was added for Torque esti-
mation delivered from a wind turbine with a sliding mode of second-order non-linear
control. The suggested overall method was evaluated using MATLAB/SIMULINK
software with a 2 MW three-blade wind turbine.
1 Introduction
Much other research on wind turbines has already been achieved. As result wind
turbines operate with variable speeds and have pitch control in recent models. We
are capable of modifying the speed and angle of pitch of each blade, Which makes
it possible for us to increase wind turbine production [1]. However, it must intro-
duce more intelligence into the operation of wind turbines. The main issues of this
article are proposed control of the g enerator which can improve the production of
the wind turbine, including energy production quality and energy efficiency. But
also Mechanical exhaustion reduction, which would create it viable to construction
lighter aero-turbines, Thus enhances productivity [1]. Therefore, control must take
the system’s behavior as a global system into account. Also, wind disturbances must
be taken into account. The studied wind turbine is a three-blade wind turbine with
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1223
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_112
1224 A. Loulijat et al.
The turbine modeling is inspired from [1–4, 7]. The disposable power of the wind
passing through an Sv surface is defined by:
1
Pwind = ρ Sv v 3 (1)
2
The wind turbine then aims to capture power following:
1
Pa = ρπ R 2 C p (λ)v 3 (2)
2
Improvement of the Non-linear Control Strategy … 1225
The power coefficient is determined by the power ratio of the wind turbine:
Pcap
Cp = (3)
Pwind
Rω
λ= (4)
v
The power captured by the wind turbine can be expressed by [1–4, 7]:
Pcap = Ta ω (5)
From the above equations, The aerodynamic torque expression can be described
by:
1 C p (λ)
Ta = ρπ R 3 Cq (λ)v 2 with Cq (λ) = (6)
2 λ
The dynamic equation of the turbine is written by [1–4, 7, 8]:
dω
J = Ta − f ω − Tem (7)
dt
The torque of reference generated by the block (1) has to answer two problems:
Maximizing wind turbine power and areas operational management [7]. The ratio
among the powers obtained from the wind and the available total power presents a
maximum known by the Betz limit. This maximum has never really occurred, And
the power coefficient of each turbine is determined by its relative speed, representing
the ratio of the speed from the tip of the turbine blades to the wind speed. Figure 2
gives an example of a power coefficient with a fixed pitch angle β for a wind turbine
changing on the tip speed ratio λ [9, 10].
More efficiently extract the wind power while keep the regime safe,the wind
turbine also must be usable following the three zones, linking wind speed, maximum
rotor speed acceptable, and the power like to see as shown in Fig. 3. In zone I,
the wind turbine is out of service because the wind is not intense to make energy
production viable about operating energy [10, 11]. In zone II, The partially loaded
wind generator works.Here the main goal is to optimize performance energy. So,
the rotation speed varies with wind speed to maintain itself at the point of operation
with maximum efficiency of aerodynamic. The aim is to ensure an always optimum
1226 A. Loulijat et al.
power coefficient [10, 11]. Zone III is a full loading activity (Intense wind). The
power must limit in order not to deteriorate the system [10, 11].
The maximum efficiency of the turbine is ensured by keeping the maximum power
coefficient [10–12]. From Eq. (6), we can write the expression of the aerodynamic
torque:
1 C p max
Ta = αω2 & α = ρπ R 5 3 (8)
2 λopt
Where the λopt specific speed helps us to maximize the power extracted [8]. The
aim of The approach used in the following for Ta to converge on Topt while Tg = Topt
is required by normal laws. This facilitation consists of ignoring The impact of
mechanical power transmission causing a power loss. A high-gain observer [13–15]
is used for estimating the aerodynamic torque for that the control is appropriate.
Compared to an observer in sliding mode [13], it allows us to limit the chattering
phenomenon in practice [14].
Improvement of the Non-linear Control Strategy … 1227
dω Ta fω Tem
= − − (9)
dt J J J
By choosing the following state variables:
Ta T
x = [x1 x2 ] = ω
T
; Y = x1 ; U = Tem
J
The state model of a high-gain observer [13–18] candidate to the system written
by the state Eq. (10) is:
X̂˙ = A X̂ + ϕ( X̂ , U ) − θ
−1 −1 −1
θ S C C( X̂ − X ) (11)
Ŷ = C X̂
Where
θ be the diagonal matrix and S is the only solution of Lyapunov algebraic
Eq. (12) [13–18]:
S + AT S + S A − C T C = 0 (12)
x̃ =
θ (x̂ − x) (13)
x̂˙ = θ (A − S −1 C T C)x̃ +
θ (ϕ(x̂) − ϕ(x)) −
θ ε(t) (14)
V (x̃) = x̃ T S x̃ (15)
V̇ (x̃) = 2 x̃ T S x̃˙ = −θ V − x̃ T C T C x̃ + 2 x̃ T S
θ (ϕ(x̂) − ϕ(x)) − 2 x̃ T S
θ ε̃(t)
(16)
Where τ = k
J
and ρ = sup t≥0 ε(t). So we have:
ρ
V̇ (x̃) ≤ −θ V + 2x̃2 σmax (S)τ + 2x̃σmax (S) (18)
θ
From the result of the mathematical Lemma 2.1. Assume that system (10) satisfies
the above assumptions. Then [18]:
2 J a 2
The sliding mode of second-order is the control law used in this part is founded on
the algorithm of super-twisting so that the estimated aerodynamic torque converges
to the optimal torque in the short amount of time [9]. Pursuit error is defined by:
Thus, we can say that there is a finite time tc for that: t ≥ tc ⇒ Ta = Topt .
The equivalent control law can obtain by annulling the Eq. (22):
s
.
The reactive power reference value will be kept at zero to maintain a unity power
factor on the stator side [24]. Within d-q coordinates, the reactive power is defined
by:
Vs φs Vs M
Qs = − Ir d (28)
Ls Ls
Vs Vs
Idr _r e f = ; φs = (29)
ωs M ωs
The proposed control strategy also uses the algorithm that says super-twisting. In
this perspective, we consider the following errors.
s Ir d = Ir d − Ir d_r e f
(30)
sTem = Tem − Tem_r e f
1 1 M M
s̈ Ir d = V̇r d + Ġ 1 − Rr Ir d & s̈Tem = − p φs V̇rq + Ġ 2 + p φs Rr Irq (32)
σ Lr σ Lr σ L s Lr σ L s Lr
The control strategy suggested in this article was validated on the 2-MW DFIG-based
WT using the MATLAB/Simulink environment. The following table illustrates the
values of some technical parameters used in the simulation (Table 1).
As can be seen in Fig. 4, we can estimate the aerodynamic torque precisely, which
allows us to control it. As concerns, the DFIG control, the performances obtained are
very satisfactory as they show the trajectory tracking of the desired electromagnetic
torque (Fig. 5a). Also, Fig. 5a shows that the torque on the shaft is not subject to the
phenomenon of chatter, thus limiting the mechanical stress on the entire transmission
of the wind turbine and Fig. 5b shows a very rapid convergence to the rotor reference
current.
To demonstrate the interest of the proposed strategy, its performances had been
comparing with control strategies such as proportional-integral control (PIC) [5]. The
PIC uses active power as a reference. This presumes that the active power conforms
to the electromagnetic power.
Ls
Ps = Pr e f ⇒ Irq_r e f = − Pr ė f (34)
Vs M
Fig. 4 Torque estimated (green), Torque optimal (blue), Torque real (red)
1232 A. Loulijat et al.
Fig. 5 Electromagnetic torque with SMSOC (a), Present efficiency of Ird tracking (b)
Ls
Irq_r e f = − Tem_r e f (35)
pMφs
6 Conclusion
A linear control method such as PIC and SVC is standard practice for A linear control
method such as PIC and SVC is standard practice for dealing with generator control
in wind turbines. However, these control methods come at the costs of decreased
system performances awaited due to random operation and process uncertainties
considered. As a consequence, a strong non-linear control method should be used
to address these issues. The method has been provided and discussed in this article.
Firstly a high-gain observer to define the reference torque in combination with the
sliding mode of second-order control, secondly a sliding mode of second-order is
selected based on a super- twisting algorithm to control the doubly-fed induction
generator, achieves the desired goals. It is strong and so durable, increases energy
efficiency, and limits mechanical stress on the drive shaft thanks to the low chatter it
produces. In future work, this strategy will be combined with the method of passive
protection in case of network disturbance in the form of a voltage dip.
References
1. Baala Y, Bri S (2020) DFIG-based wind turbine control using high-gain observer. In: 2020 1st
Interanational Conference Innovation Research Application Science Enginnering Technology,
IRASET 2020 (2) https://doi.org/10.1109/IRASET48871.2020.9092280
2. Goudarzi N, Zhu WD (2013) A review on the development of wind turbine generators across
the world. Int J Dyn Control 1(2):192–202. https://doi.org/10.1007/s40435-013-0016-y
3. Lin FJ, Tan KH, Fang DY (2015) Squirrel-cage induction generator system using hybrid wavelet
fuzzy neural network control for wind power applications. Neural Comput Appl 26(4):911–928.
https://doi.org/10.1007/s00521-014-1759-x
4. Mustafaev RI, Gasanova LG (2015) An investigation of the dynamics of a wind power unit
equipped with synchronous generators with permanent magnets. Russ Electr Eng 86(5):258–
263. https://doi.org/10.3103/S1068371215050077
5. Poitiers F, Bouaouiche T, Machmoum M (2009) Advanced control of a doubly-fed induction
generator for wind energy conversion. Electr Power Syst Res 79(7):1085–1096. https://doi.org/
10.1016/j.epsr.2009.01.007
6. Cárdenas R, Peña R (2004) Sensorless vector control of induction machines for variable-speed
wind energy applications. IEEE Trans Energy Convers 19(1):196–205. https://doi.org/10.1109/
TEC.2003.821863
7. Beltran B, Ahmed-ali T, El M, Benbouzid H, Member S (2008) Sliding mode power control
of variable-speed wind energy conversion systems 23(2):551–558 (2008)
8. Touati A, Abdelmounim E, Aboulfatah M, Moutabir A, Majdoul R (2015) Improved strategy
of an MPPT based on the torque estimator for variable speed turbines. Int Rev Model Simul
8(6):620–631. https://doi.org/10.15866/iremos.v8i6.7122
9. Abd I, Abdelmounim E, Aboulfatah M, Majdoul R, Moutabir A (2015) Design of an Mppt
based on the torque estimator for variable speed turbines (1):166–171
10. Beltran B, Benbouzid MEH, Ahmed-Ali T (2020) A combined high gain observer and high-
order sliding mode controller for a DFIG-based wind turbine. In: 2010 IEEE International
Energy Conference Exhibilish EnergyCon 2010 (1):322–327 (2010). https://doi.org/10.1109/
ENERGYCON.2010.5771699
1234 A. Loulijat et al.
1 Introduction
Due to the fact that the real nature of wind speed is stochastic, the grid-connected
DFIG is provided with automatic control strategies for optimal power extraction and
efficient exchange of energy between the turbine and the grid.
The aim of the MPPT strategy, when the turbine is on partial load is to continually
adjust the operating point of the wind turbine in order to convert the maximum power
possible. A technique using relative velocity [1] has been proposed in the literature.
Meanwhile, wind speed acquisition from a single anemometer leads to erroneous
measure of wind speed which does not the effective wind speed incident on the
blades. Thus, a faulty measurement of the incident wind speed can cause the deterio-
ration of the extracted power. Wind speed estimation with two different Kalman filter
versions were proposed and analyzed by the authors in [7]. The complexity of Kalman
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1235
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_113
1236 H. Chojaa et al.
filters and the high number of tuning parameters required are the major drawbacks
of this method. A simple method based on Fuzzy Logic were synthesized in [3] to
estimate wind speed and track the DFIG variables. Easy to implement Perturbation
and Observation technique usually employed in photovoltaic systems is proposed in
[5]. It can be implemented easily, not expensive and does not require accurate param-
eters of the turbine and wind speed information. However, a compromise between
convergence speed and performance is needed. Neural Network, Gradient methods
have been proposed to give very good performance but are yet to be easy to imple-
ment [1]. The disadvantage of the Optimal Torque Control technique presented in
[9] is that characterization tests are required off-line to design the look-up table. This
paper proposes a scheme based on Fuzzy Logic, which is unrestrained by wind speed
measurement, knowledge of turbine parameters and characterization experiment to
search the maximum power point. It indirectly compensates for the different power
loss and requires only the acquisition of DFIG speed and output power signals.
Vector control schemes and PI-controllers have been employed hand-in-hand in
the literature to regulate DFIG velocity, power, and currents. However, grid voltage
dips, model incorrectness and unexpected conditions like temperature increase dete-
riorate the effectiveness of this control technique. To improve on these inconve-
niences, several other controllers like Sliding Mode [9, 10], Backstepping [2, 8],
Feedback Linearization [6], Neural Predictive Control [4] have been studied and
evaluated in the literature. The drawbacks of these approaches range from chattering
to implementation complexity.
In this paper, we design an adaptive control technique in line with Lyapunov
stability theorem to regulate speed and current. Estimating and compensating uncer-
tainty in real time, it is characterized by fast response time, good precision, better
robustness against disturbances.
dir d vr d Rr ir d L m ωr ϕsq
= − + ωr irq + + θr d (1)
dt σ Lr σ Lr σ Lr Ls
dirq vrq Rr irq
= − − ωr ir d + θrq (2)
dt σ Lr σ Lr
dωg 3 Lm
J = Tm − f v ωg − p ϕsq ir d + θω (3)
dt 2 Ls
Ps = − 23 LLms vsd ir d
(4)
Q s = − 23 vLsds ϕsq − L m irq
3 Controller Design
Three controllers are designed in for the work presented in this paper: pitch angle,
RSC and GSC controllers. The pitch angle controller has for objective to mechan-
ically prevent blade rotation speed from exceeding nominal power and speed. The
aim of the RSC controller is to adaptively drive and set the system on the maximum
power point and track desired stator reactive power. The PI corrector, designed but
not presented in this paper to control the GSC, is aimed at maintaining DC voltage
constant and annulling filter reactive power.
1238 H. Chojaa et al.
Figure 2 presents the curve of mechanical power in function of generator speed for
different wind speed values. Popt represents the maximum power point (MPP) of the
turbine system. Since the wind speed is random and not fix, to obtain the optimum
output power of the wind turbine, an adaptive controller is needed to track the optimal
power curve through the MPPT strategy.
The MPPT strategy in this paper adopts the fuzzy technique described in Fig. 2(b),
well known as an adaptive and robust tool for complex nonlinear systems. It is
employed in this paper because it neither needs the information of turbine param-
eters nor wind speed signal acquisition. The FLC perturbs the generator speed and
observes the direction of variation of the output power and adjusts the speed accord-
ingly in other to place the turbine at the operating point where energy extracted is
maximal. The output is the optimal generator speed (ωopt ) that ensures the extraction
of maximum power from the turbine blades. The speed reference is then fed to a fast
controller which converges the real generator speed to the reference. The calculation
of output power from stator and filter measurement is more efficient because it avoids
the complex determination of power loss in the gearbox, the converters and the filter.
The fuzzy logic controller synoptic diagram is shown in Fig. 3, where K1, K2 and
K3 are adaptive gains.
The FLC searches the maximum point of the power-speed curve of the generator
using the rules presented in Table 1: with P and ωg as inputs, while ωgr e f
represent the output. These variables are represented by NB, NM, NS, Z, PS, PM,
PB denoting Negative Big, Negative Medium, Negative Small, Zero, Positive Small,
Positive Medium, and Positive Big respectively.
The triangular membership functions of input and output are presented in Fig. 4
and have seven fuzzy subsets. The fuzzy inference is carried out by using Sugeno’s
method and the defuzzification uses the center of gravity to determine the output of
this FLC.
The most important feature of the ABC is that, unlike the PIC, it takes eventual uncer-
tainty (θω , θr d , θrq ) into account in the modelling, estimates it and sends commands to
compensates for it in real time. This is essential to mitigate the effect of model incor-
rectness, parameter variation, modelling approximations, and saturation. Current
reference value which guarantees the convergence of generator speed to its optimum
is derived in step 1. An ABC is designed to tracks this reference current in step 2 to
give the final control variable vr d .
Step 1: Eqs. 5 and (6) show speed tracking error and its integral respectively.
εω = ωg − ωopt (5)
Tm f v ωg 3 Lm
ε̇ω = − − p ϕsq ir d + θω − ω̇opt (6)
J J 2 J Ls
The current output ir∗d is the reference input for step 2 and is derived from (8) and
presented in (9). Equation 10 represents the estimate of the uncertainty computed in
real time.
2J L s Tm f
θ ω = εω m ω (10)
Step 2: The control of stator reactive power will be included to this step. The
∗
error between the actual and the desired current can be defined as (irq = α2 ):
Optimization of DFIG Wind Turbine Power Quality … 1241
εr d = ir d − αr d
(11)
εrq = irq − αrq
L ω ϕ
ε̇r d = σvrLdr − RσrLirrd + ωr irq + σmL rrL ssq + θr d − α̇r d
v R i (12)
ε̇rq = σ rqL r − σrLrqr − ωr ir d + θrq − α̇rq
⎧ 2
⎨ Vr d = Vω + 1 ε2 + 1 θ r d
2 rd 2 mr d
2 (13)
⎩ V = 1 ε2 + 1 θ rq
rq 2 rq 2 m rq
Second step Lyapunov functions taking into account uncertainty parameters and
their derivatives, are presented as (13) and (14) respectively.
V̇ω = −K ω εω2 − K r d εr2d
(14)
V̇rq = −K rq εrq
2
Derivatives of the estimates of the uncertainties and the control variables deduced
and given as (15) and (16) ensure that the derivative of the Lyapunov function is
strictly negative.
˙
θ r d = εr d m r d
˙
(15)
θ rq = m r d εr d
⎧
⎨ vr d = σ L r −K r d εr d + 3 p εω L m ϕsq + Rr ir d + ωr irq − ωr L m ϕsq − θ r d + α̇r d
2 J L s σ L r σ L L
r s
⎩
R i
vrq = σ L r −K rq εrq + σrLrqr + ωr ir d − θ rq + α̇rq
(16)
The DFIG model studied in this work is rated 1.5 MW. It is designed under
Matlab/Simulink to examine the performance of the ABC in comparison to that
of the PI-controller taking the stochastic wind speed profile presented in Fig. 6(a)
as input. Figure 6(b) shows that the PIC maintains the DC bus voltage constant for
effective power conversion. Particular interest is placed on the convergence time,
precision, and power quality under pitch angle control activation.
Figure 6(c) proves that actual speeds converge to corresponding reference gener-
ator speeds under the ABC and PIC with good precision and robustness. We however
observe discrepancies between speed evolution under both controllers. This is due
1242 H. Chojaa et al.
(a) (b)
(c) (d)
(e) (f)
(g) (h)
Fig. 6 a Wind speed b DC-link voltage c Generator speed d Power coefficient e Electromagnetic
Torque f Active and reactive power at the point of common coupling g Stator current under ABC
h Stator current under PIC
principally to the better precision and faster convergence of the ABC. This can be
explained by the higher power coefficient obtained under the ABC as illustrated in
Fig. 6(d).
The rapidity of the ABC enables the FLC to take optimum decision more quickly,
leading to better precision and more extracted power. It can also be noticed on
Optimization of DFIG Wind Turbine Power Quality … 1243
Fig. 6(e) that the electromagnetic torque is less noisy under the ABC. This conse-
quently implies less mechanical vibration and better power quality. Figure 6(f) shows
that the powers transmitted to the grid under both controllers possess smooth evolu-
tion and the reactive power converges to zero with more power extracted under
the ABC without chattering. The stator current waveforms under ABC in Fig. 6(g)
takes a better sinusoidal form than that of the PIC given in et Fig. 6(h). This conse-
quently demonstrates the better performance of the ABC. Due to the limitation of
the reference speed below 1.2 p.u, the pitch angle remains close to zero for both
controllers.
5 Conclusion
Optimizing energy conversion in DFIG wind turbine requires to, first, determine
the Maximum Power Point (MPP) and second, control rotor converters to set the
system on the MPP. Due to uncertainties and model imperfection, a Fuzzy Logic
adaptive controller is employed in this paper to track the optimum generator speed.
Comparative studies and simulation results show that the Adaptive Backstepping
rotor converter controller proved more power conversion, better robustness, faster
response time and higher precision than the conventional PI controller.
References
1. Abdullah MA, Yatim AHM, Tan CW, Saidur R (2012) A review of maximum power point
tracking algorithms for wind energy systems. Renew Sustain Energy Rev 16:3220–3227.
https://doi.org/10.1016/j.rser.2012.02.016
2. Adekanle OS, Guisser M, Abdelmounim E, Aboulfatah M (2017) Robust integral backstepping
control of doubly fed induction generator under parameter variation. Int Rev Autom Control
10. https://doi.org/10.15866/ireaco.v10i2.11163
3. Belmokhtar K, Doumbia ML, Agbossou K (2014) Novel fuzzy logic based sensorless maximum
power point tracking strategy for wind turbine systems driven DFIG (doubly-fed induction
generator). Energy 76:679–693. https://doi.org/10.1016/j.energy.2014.08.066
4. Douiri MR, Essadki A, Cherkaoui M (2018) Neural networks for stable control of nonlinear
DFIG in wind power systems. Procedia Comput Sci 127:454–463. https://doi.org/10.1016/j.
procs.2018.01.143
5. Hong C-M, Ou T-C, Lu K-H (2013) Development of intelligent MPPT (maximum power point
tracking) control for a grid-connected hybrid power generation system. Energy 50:270–279.
https://doi.org/10.1016/j.energy.2012.12.017
6. Li P, Xiong L, Wu F, Ma M, Wang J (2019) Sliding mode controller based on feedback lineariza-
tion for damping of sub-synchronous control interaction in DFIG-based wind power plants. Int
J Electr Power Energy Syst 107:239–250. https://doi.org/10.1016/j.ijepes.2018.11.020
7. Song D, Yang J, Dong M, Joo YH (2017) Kalman filter-based wind speed estimation for wind
turbine control. Int J Control Autom Syst 15:1089–1096. https://doi.org/10.1007/s12555-016-
0537-1
8. Youness EM, Aziz D, Abdelaziz EG, Jamal B, Najib EO, Othmane Z, Khalid M, Bossoufi
B (2019) Implementation and validation of backstepping control for PMSG wind turbine
1244 H. Chojaa et al.
Abstract Wind energy is considered one of the most important and promising
renewable energy sources in the world. We begin this study with a description of
the current state of the world energy situation and in Morocco in particular. We
then described a medium power wind power conversion chain that we modeled
and simulated. The chain consists of the turbine, the multiplier and the dual
power induction generator (DFIG). The adopted chain model is developed using
Matlab/Simulink/Sim-Power-Systems software. We found fairly correct simulation
results which are presented and analyzed at the end of this study.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1245
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_114
1246 W. Aicha et al.
energy, which encouraged several countries like Morocco to develop other forms of
energy: renewable energies.
In the conversion of wind energy, the DFIG is widely used because it has several
ad-vantages: the generation of energy at variable speed (±30% around the speed of
synchronism) which allows to obtain the maximum possible power for each wind
speed: MPPT. The wind chain, as shown in the following figure, includes the turbine,
the multiplier, the DFIG whose stator circuit is directly connected to the electrical
network. A second circuit placed at the rotor is also connected to the network, but via
power converters. The machine side power converter is called “Rotor Side Converter”
(RSC) and the line side power converter is called “Grid Side Converter” (GSC).
The machine-side power converter is used to control the active and reactive power
produced by the machine. The mains side converter controls the DC bus voltage and
the mains side power factor (Fig. 3).
1248 W. Aicha et al.
d∅ds
vds = Rs ids + − ωs ∅qs (1)
dt
d∅qs
vqs = Rs iqs + ωs ∅ds (2)
dt
d∅dr
vdr = Rr idr + − ωr ∅qr (3)
dt
d∅qr
vqr = Rr iqr + + ωr ∅dr (4)
dt
⎧
⎪
⎪ ∅ds = Ls ids + Midr
⎨
∅qs = Ls iqs + Miqr
(5)
⎪ ∅dr
⎪ = Lr idr + Mids
⎩
∅qr = Lr iqr + Miqs
After having studied each element separately, we will now be interested in the system
in its totality: model it and simulate it. The complete open loop system can be
represented by the following block diagram (Fig. 4).
Use of a Doubly-Fed Induction Generator … 1249
Note that the three-phase system of currents (is1, is2, is3) is balanced with the
same pulsation as the voltages. Imax ≈ 6 A.
The Fig. 8 shows the total active power supplied by the generator.
Use of a Doubly-Fed Induction Generator … 1251
Fig. 8 Total active power supplied by the wind chain to the load
It can be seen that the wind chain provides a fairly high active power: average P =
2000 W. (The minus sign indicates of course that it is a supplied power). We note the
presence of fluctuations (of the same frequency as the voltage) on the power plot which
are due to the use of power converters and to the fact that we worked without power
regulation.
The last figure shows the total reactive power supplied by the chain (Fig. 9).
The average value of the reactive power is zero since the load we have placed is
of a resistive nature.
1252 W. Aicha et al.
In the light of the results given previously, we can conclude that the studied conversion
chain was modeled correctly. The simulation of the design carried out using the
software are quite valid: the currents, the voltages as well as the delivered powers
correspond to the desired results.
In order to minimize the observed parasitic fluctuations, we plan to insert filters
at the terminals of the load.
References
1. Boyette A (2006) Control command of a dual power asynchronous generator with storage system.
Doctoral thesis from Poincaré University, Nancy
2. Wahabi A, El Moudden A, Bounifli F (2014) Modeling and Simulation of a Dual Power
Asynchronous Generator for Wind Energy Control. Moroccan thermal association (2014)
3. Belmokhtar J, Doumbia K, Agbossou K (2010) Modeling and control of a wind power system
based on a dual power asynchronous machine for the supply of power to the electricity grid. J
Sci Res 2
4. Khojet el khil S (2006) Vector control of a double-fed asynchronous machine
(MADA)/Optimization of losses in converters 2006
Use of a Doubly-Fed Induction Generator … 1253
Abstract Given the role of PMSG in the wind power system, several commands
are proposed to improve the quality of energies supplied to the electrical grid. In this
work, two control strategies are presented and applied to the static converter in order
to control the electromagnetic torque and the active and reactive power exchanged
between the PMSG and grid. Firstly, in order to extract a maximum power from
the wind a speed control strategy (MPPT) has been presented, followed by a PMSG
model in reference d-q. Then, the conception of the Field-Oriented Control (FOC)
and the sliding mode control (SMC) are developed and tested with Matlab/Similink,
where these controls make it possible to operate the system in the best performances
in dynamic regimes. And finally, we ended by a discussion about the interpretation
of the simulation results in order to compare the advantages and the performances
of each control.
1 Introduction
Recently, the development and the exploitation of renewable energy have experi-
enced a strong progress, due to its different advantages (technical, economical and
environmental) compared to conventional energies, including fossil energies (coal,
gas, petrol). There are many types of renewable energy resources (wind, solar, hydro,
geothermal and biomass) [1–3]. Where the wind energy occupies a good place, it is
one of the cleanest, the safest and fastest sectors to implement, it is supposed to be
the best in terms of quality and price [4–6].
There are several research studies about the wind turbine. In particular, those
equipped with Permanent magnet synchronous machine. This type of machine is
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1255
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_115
1256 B. Majout et al.
1
Pwind = .ρ.S.vw3 (1)
2
1
Pturb = Cp (A, β).Pwind = .ϕ.S.Cp (A, β).v3 (2)
2
Improvement of the Wind Power System Based on the PMSG 1257
t .R
λ= (3)
v
16
p (λ, β) =
Cmax = 0.5930 (4)
27
−C5
C p (A, β) = C1 CA2 − C3 β − C4 e A + C6 .λ
(5)
1
A
= λ+0.08.β
1
− 1+β
0.035
3
Ptur b 1 1
Ttur b = = .ϕ.S.C p (A, β).v 3 . (6)
t 2 t
dmec
Jtot = Tmec = Ttur b − Tem − f.mec (8)
dt
The parameters C1, C2, C3, C4, C5 and C6 are constants dependent on the turbine
geometry (Table 1) [7].
According to Fig. 2, the plot of the Cp follows the variation of the Beta angle
inversely and reaches its maximum 0.426, for λopt = 8 and B = 0°. Given that the
wind turbine can’t convert more than 59% of the kinetic energy contained in the wind
into mechanical energy according to Betz’s Law [7–9].
Fig. 2 Curve of Cp
1258 B. Majout et al.
The electrical equations of the PMSG in q-d reference are as follows (see Fig. 4) [10,
11]:
dsd
V sd = Rs .i sd + − wr .sq (10)
dt
dsq
V sq = Rs .i sq + + wr .sd (11)
dt
wr = p.mec (13)
3
Tem = . p L d − L q Isd .Isq + Isq .∅ f (14)
2
The active and reactive power are expressed by:
Pgen = 23 Vsd.isd + Vsq.isq
(15)
Qgen = 23 Vsq.isd − Vsd.isq
3 Synthesis of Controls
In order to guarantee a high performance of a wind energy system based on the PMSG
with smooth poles, where the inductances Ld and Lq are equal L s = L d = L q , The
frequency and the voltage of the generator connected to the distribution grid must be
adjusted when the wind speed varies [7].
We use the following PMSG model [7–12].
d Isd 1
I˙ds = =− Rs .i sd − p.mec .L q .i sq − V sd (16)
dt Ld
d Isq 1
I˙qs = =− Rs .i sq + p.mec .L d .i sd + p.mec .∅ f − V sq (17)
dt Lq
1 3 p f
˙ mec =
.Ttur b − . . L d − L q Isd .Isq + Isq .∅ f − .mec (18)
Jtot 2 Jtot Jtot
The PMSG model represented by differential Eqs. (10) and (11) indicate that Id
and Iq are coupled together by nonlinear terms mec .L d .i sd and mec .L dq .i sq [7].
To solve this problem, two controls have been developed in this paper in order to
control the machine-side converter:
1) Field-Oriented Control (FOC).
2) Sliding Mode Control (SMC).
1260 B. Majout et al.
The basic principle of vector control was described by BALSCHKE and HASSE in
the early ’70s [10–16]. the Vector control consists in eliminating the coupling between
the inductor and the armature by decomposing the stator current into two parts (Isd,
Isq), where the inductive current controls the magnetic flux and the armature current
controls the electromagnetic torque. The objective of this command is to establish
a simple model of PMSG which makes it similar to a DC machine with separate
excitation [13, 14]. The vector control is generally based on Proportional-Integral
regulators PI to regulate the error between the measured input and desired output.
According to the compensation method, the expression of the reference voltages
after removing the terms of compensation −wr .Lq .isq and wr (Ld .isd + ∅f ) become:
[13–17]:
V sdref = (Rs + S.Ld )isd
(19)
V sqref = (Rs + S.Lq )isq
In order to decouple the electromagnetic torque and control it by Isq only, the
vector control imposes Id_ref = 0. the expression of the electromagnetic torque
becomes as follows [7, 13].
3 p
Tem = . .Isq .∅f (20)
2 Jtot
The sliding mode control technique considered one of the special cases of the VSS
variable structure system, which is characterized by a choice of the structure and
the switching logic [16]. The evolution of this system depends only on properties
of the sliding surface. The principle of this method consists in forcing the system
state trajectory to attain a hyper surface in finite time and then stay there, where the
behavior of results corresponds to the desired dynamics [14–18].
The implementation of this method of control requires three complementary steps
concerning [16–18]:
• Selecting the sliding surface.
• Defining the convergence conditions based on Lyapunov functions.
• Determining the strategy of the control.
With,
e(x): the error in the output state e(x) = xref – x.
δ: a positive coefficient.
r: the rth order system. Where for r = 1, S(X) = e(x).
For the order system r = 1, the sliding surface for the dq-axis current and
mechanical speed generator can be presented as follows [7–18]:
S Isq = e(Isq ) = Isq ref − Isq (24)
f 3 p 1
˙ mec−ref +
Ṡ(mec ) = .mec + . .Isq .∅ f − .Ttur b (25)
Jtot 2 Jtot Jtot
Rs Lq V sd
Ṡ(Isd ) = I˙sd −r e f + .i sd − p.mec . .i sq − (26)
Ld Ld Ld
Rs Ld ∅f V sq
Ṡ Isq = I˙sq −r e f + .i sq + p.mec . .i sd + p.mec . − (27)
Lq Lq Lq Lq
The sliding mode occurs on the sliding surface when Ṡ(x) = 0 and Vsd,q _N = 0
[16–18], We can deduce the expression of the equivalent command Veq:
Rs Lq
Vsd _eq = L d I˙sd −r e f + .i sd − p.mec . .i sq (29)
Ld Ld
Rs Ld ∅f
Vsq _eq = L q I˙sq −r e f + .i sq + p.mec . .i sd + p.mec . (30)
Lq Lq Lq
Vsq _N = K q .sat S Isq with K q > 0 (32)
Speed Regulator
For the speed controller, the control variable becomes the reference current of the
quadrature axis.by Following the same method used for controlling the d and q axes,
the expression of Isq _ref is given as follows [7–18].
Isq _eq = − 23 . p.∅
Jtot
˙ mec−r e f + f
.mec − 1
.Ttur b
f Jtot Jtot (34)
Isq _N = K mec .sat(S(mec )) with K mec > 0
In order to evaluate the performance of both control strategies FOC and SMC, we
tested the behavior of the wind system based on PMSG in a dynamic regime, by
a test performed in MATLAB/Simulink, where the turbine is trained by a variable
wind speed using the MPPT technique to approach reality (see Fig. 6). The wind
system parameters (Turbine and PMSG) are given in Table 3.
The following figure illustrate the results of the PMSG simulation in the Matlab
/Simulink environment:
From curves shown in Fig. 7, we can notice that the sliding mode control offers
better results compared to vector control based on the classical PI regulators. On the
Improvement of the Wind Power System Based on the PMSG 1263
other hand, the references fixed for both controls are well followed by the electro-
magnetic torque and the active and reactive powers as well as the stator currents. In
addition, the reactive stator power is null which gives a unit power factor.
The electromagnetic torque varies according to the variation of the mechanical
torque. The torque ripple for SMC control is strongly reduced compared to the torque
ripple for FOC control.
The control of the rotation speed is well ensured. The Zoom 1 of the rotational
speed shows clearly the quality of the SMC control compared to FOC on the mechan-
ical rotational speed as well as the rapidity and precision. We note also that the static
and dynamic errors are infinitely low in the case of SMC.
The good quality of the stator currents in the rotating frame of reference gener-
ates sinusoidal three-phase currents in good quality. However, the frequency of the
currents changes with the change of the mechanical rotation speed which is propor-
tional to the variation of the wind speed. The ripple rate of the currents for SMC
control is low than the rate found in the case of vector control.
The Table 2 shows the time when each response starts to follow its reference.
5 Conclusion
This paper presents a comparative analysis between SMC and FOC control applied
to the machine-side static converter for a variable speed wind turbine based on the
permanent magnet synchronous generator. In the first time, We presented the vector
control then the non-linear SMC control in the second time. According to the results
shown in Fig. 7 and Table 2, we notice that the vector control offers an acceptable
result, whereas the SMC control offers a high static and dynamic performances.
Therefore, the SMC control remains the most appropriate solution to be applied for
the control of the three-phase rectifier.
1264 B. Majout et al.
Fig. 7 Simulation results of the Field-Oriented Control and Sliding mode control
Improvement of the Wind Power System Based on the PMSG 1265
Table 2 Summarization
FOC (Tr) (ms) SMC (Tr) (ms)
table of the response time
SMC and FOC Controller 20 ms 10 ms
Ps 20 ms 12 ms
Qs 22 ms 16 ms
Tem 50 ms 30 ms
Table 4 Nomenclature
Nomenclature
Pturb Power captured by the wind turbine S Blade swept area
Vw Wind speed R Radius of the turbine blade
mec Mechanical generator speed ωr Electric pulsation
t Mechanical turbine speed Pgen Active generator power
Tem Electromagnetic generator torque Qgen Reactive generator power
Tem_ref Reference generator torque Vs Stator voltage vector
Ttur Turbine torque Rs Stator resistance
Cp(λ, β) Power coefficient L(d,q) d-q axis inductance
λ Tip-speed ratio Vs(d,q) d-q axis stator voltage
β Pitch angle is(d,q) d-q axis stator current
ρ Specific density of air (d,q) d-q axis flux
p Number of poles pairs ∅f Generator flux
References
1. Ackermann T, Soder L (2002) An overview of wind energy-status 2002. Renew Sustain Energy
Rev 6(1e2):67e127
1266 B. Majout et al.
Abstract To solve the problem of water for irrigation in isolated regions and rural
remote areas, where the expansion of the power network line is complex and she is
more costly. In this paper a control of the wind pumping systems is presented. Clas-
sical control strategy of induction motors (IMs) is very sensitive to the variation of
rotor resistance, to leads a strong coupling between the rotor flux and the electromag-
netic torque. To resolve this difficulty, we propose in this paper, a nonlinear control
of high performance applied of induction motor, to improve the tracking control, for
guarantee stability, robustness to parameter variations, perfect decoupling between
the flux and torque. The robustness of the proposed study is confirmed through
software simulation with parameter variations.
1 Introduction
Water supply, livestock and domestic necessities are main issues in the progress of
isolated and Saharan regions, where the extension of the conventional power network
remains very hard to do [1, 2]. Therefore, recently it is necessary to use inexhaustible
resources to fulfill their energy requirements. [2, 3].
The demand for rural water for domestic supplies and irrigation is increasing, with
the result that surface water become increasingly scarce in the future [4, 8]. For this,
it seems that the only and good alternative to this problem is groundwater. but the
water table also decreases, making manual pumping more challenging. However, the
pumping of water mechanized will be the only dependable substitute for lifting water
A. Loubna (B)
Department of Electrical Engineering, Oum El Bouaghi University, Oum El Bouaghi, Algeria
T. Riad · M. Salima
Department of Electrical Engineering, Laboratory of Electrical and Renewable Energies LEER,
Souk Ahras University, Souk Ahras, Algeria
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1267
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_116
1268 A. Loubna et al.
from the ground. Even so, water pumping in the world usually depends on conven-
tional electricity produced by electricity or diesel. A photovoltaic and wind pumping
system has appeared as a satisfactory alternative to classical pumping systems used
in isolated regions and rural remote areas. Pumping water through wind energy mini-
mizes costly fuels, noise and air pollution [9]. The water pumping systems with wind
turbines are generally composed of a wind turbine, an electric generator, one or more
converters and a motor generally driven the centrifugal pump load [7].
A squirrel cage induction generator (SCIG), and DC Motor the Permanent Magnet
Synchronous Generator (PMSG), were generally most used in the water pumping
[7–10]. the permanent magnet synchronous generator (PMSG) has been gaining
popularity due to its high mass torque efficiency with low inertia ratio, its simple
of control, recently are good competitor of the induction motor in a wind energy
conversion [11, 12].
For this, the research has oriented to other solutions involving less expensive and
more robust actuators, namely the Induction machine, more reliable and less costly
in terms of construction and maintenance [14, 15].
Induction machine is careful as a distinctive actuator in the all industries for vari-
able speed applications since it has many advantages such as reliability, simplicity
of design, high efficiency, good self-starting, the absence of the collector brooms
system, low maintenance, and relatively reasonable budget. [2, 16], However, it
dynamic model is nonlinear and there control is very complex, due the high coupling
between the input variables, output and internal variables [16–19]. But the advance-
ment of power electronics, made a fundamental revolution which its m main objective
is to develop the recent control strategies for IM. Several control strategies have been
developed and used for IM drives such as: Direct Torque and Flux Control (DTFC),
Vector Control (VC) [13, 16].
However, this technique demands the use of a magnetic sensor based on Hall
Effect, to measure the magnitude of the rotor flux, this sensor remains very sensitive
to motor parameters variations and raises the cost of the drive equipment [16, 21, 22].
To solve this problem involved in a vector control, many researchers presented
new control strategy using the rotor flux and speed estimation without any sensor
in order to decrease the expenses and have the features of precision, elimination of
disturbance and fast torque response [22]. One among nonlinear control methods is
the nonlinear control with input output linearization, based on the transformation
of the nonlinear model into several linear system; which makes the linear feedback
control easy to be applied in respect to trajectory reference [16, 22, 23], The input
output linearization control based on differential geometry theory is used successfully
to eliminated coupling problem, resulting in a rapid transient response with decoupled
flux and torque response [16]. In recent decades, the theory of feedback control of
nonlinear seen significant developments in various industrial areas and has now
the axis of several research status, it is based on theoretical concepts of differential
geometry such that the Lie derivative and deffeomorphism, it allows diffeomorphic by
a nonlinear transformation and an uncoupled state feedback to linearize the template
and placed in the canonical form [6, 19]. In this background several research paper
have been published in various research laboratory.
Control of Wind Water Pumping … 1269
In [25] the authors present a control technique of water pumping system based on
wind turbine generator in an isolated area (in South of Algeria), Simulation results
have confirmed satisfactory performance and verified the efficiency of the suggested
pumping system. In [26] The authors provided control of a water pumping system
based on a wind turbine generator in an isolated area by using the Maximum Power
Tracking Point (MPPT), The wind energy conversion system is used to drive the
permanent magnet synchronous generator (PMSG) to feed the isolated load which
is consisted of DC motor drives a hydraulic centrifuge pump. In [27] The authors
applied the vector control of a wind power conversion system based on the doubly fed
induction machine (DFIM) connected to the grid. In [28] presented an experimental
study on the efficient field oriented control with power losses for the optimization
of the energy production of a wind energy conversion chain based on the six phase
asynchronous generator. In [29] the authors used the scalar command for controlling
a PV water pumping system, consists in driving the machine with a constant flow for
variable voltages, the simulation results showed the effectiveness and feasibility of
the suggested method. In [2] the authors presented an robust control technique based
input-output linearization control of IM coupled with hydraulic pump powered by a
PV panel. Simulation results have confirmed good performance of the proposed PV
Water pumping system.
To solve the problems of high coupling between the flux and torque in vector
control strategy and to improve the tracking control, for guarantee stability, robust-
ness to parameter variations, perfect decoupling between the flux and torque, we
propose in this paper, a nonlinear control of high performance applied of (IM). In
the proposed system, the (PMSG) is wind turbine driven. The centrifugal pump is
driven by an IM and fed by a voltage-source inverter (VSI).This article is presented as
follows; the description of wind pumping system is evolved in Sect. 2. In Sect. 3 feed-
back linearization control of IM will be presented. In Sect. 4 the performance of the
proposed control of the water pumping system is verified by numerical simulation.
Finally, some conclusions are given in Sect. 5.
The robustness of the proposed control system is verified through numerical
simulation with motor parameters variations (Figs. 2, 3, 4, 5, 6 and 7).
The wind turbine captures the kinetic energy of the wind, passing through the section
of the active surface swept by the turbine blades, and converts it into mechanical
energy (torque) which turns the rotor of an electric generator [26, 30]. The mechanical
power of the wind turbine is given by the following expression [4, 31]:
1270 A. Loubna et al.
1
Ptur = ρ SV 3 C p (λ) (1)
2
Where: Ptur : is the extracted power, ρ: is the air density [kg/m3 ], V: is the wind
speed [m/s], S: is the area swept by the rotor blades, C p : is called the power coefficient.
λ is tip speed ratio of rotor blade which equals a tip speed to a wind speed is
written as follows [32, 33].
t R
λ= (2)
V
R: is the radius, t : is the mechanical speed of the wind turbine rotor.
In the (d, q) axis the model of the PMSG used in the modeling model is given by the
following equations [4, 11, 26]:
ψd Ld 0 id ψv
= + (3)
ψq 0 Lq iq 0
Vd = Rs .i d + dtd ψd + ωr ψq
(4)
Vq = Rs .i q + dtd ψq − ωr ψd
Control of Wind Water Pumping … 1271
dr
Tem − Tr = J + f.r (5)
dt
The equation of electromagnetic torque is expressed by (6) [11]:
3p
Tem = ψ f iq + L d − L q id iq (6)
2
ψ f : Rotor Flux.
• Model of IM. The model of the IM in the stationary frame (α, β) is given by (7)
[2, 21, 34]:
x = f (x) + g(x).u
(7)
y1 = h 1 (x); y2 = h 2 (x)
Where
⎡ ⎤
−γ Isα + TKr φr α + pK φrβ
⎢ −γ Isβ − pK φr α + T r φrβ
K ⎥ T t
⎢ ⎥ 1
⎢ ⎥ g1 0 000
f (x) = ⎢ T r Isα − T r φr α − pφrβ
M 1
⎥ [g] = = σ LS
⎢ ⎥ g2 0 1
σ LS 0 0 0
⎣ T r Isβ + pφr α − T r φrβ
M 1 ⎦
p J L r (φr α Isβ − φrβ Isα ) − J (TL + f )
M 1
x = [Isα Isβ φr α φrβ ]T : u = [u sα u sβ ]T , y = (φr2α + φrβ
2
)
Lr M2 M Rs Rr M 2
Tr = ; σ =1− ;K = ;γ = +
Rr Ls Lr σ Ls Lr σ Ls σ L s L r2
PM
Tem = (φr d Isq − φrq Ir d ) (8)
Lr
dr
J = Tem − TL − f r (9)
dt
1272 A. Loubna et al.
H pump = k0 + k1 Q + k2 Q 2 (10)
Where: k 0 , k 1 , k 2 : are the constant of the pump, H[m]: the head of the pump and
Q [L/s]: the flow rate of the pump [35, 36].The IM is coupled to centrifugal pump,
by the following mechanical equation [2, 2]:
d
J = Tem − T pump (11)
dt
The torque- speed relation of this pump is given by.
T pump = k T 0 + k T 1 + k T 2 2 (12)
Where, k T0 , k T1 , k T2 are the empirical constants of the pump which can be obtained
from the measurements.
The difficult case for the principle of the linearization input-output decoupling is
the choice of output variable yi. For the application of linearization methods state
feedback to control the asynchronous machine, there are two cases where we can
[15, 16]. The system to be controlled by a control law linearization must be square
type. As a result, we chose to rotor speed and magnitude of rotor flux as outputs of
the process [15].
⎡ ⎤
1 2 2
h 1 (x) φ
⎢ 2 rα + φ 2
rβ ⎥ φr α + φrβ
2
y= =⎢
⎣ pM
⎥
⎦ = (13)
h 2 (x) Ce
Isβ φr α − Isα φrβ
Lr
h 1 (x)
y= = 2 (14)
h 2 (x) φr α + φrβ2
To obtain the optimal non-linear law control, we determine the relative degree
of the output y1 and y2 , that is to say the number of times to derive the output to
show the input Vsα Vsβ .changes in output are selected given by the Eq. (14) such that:
[15, 15].
Control of Wind Water Pumping … 1273
⎧
⎨ h 1 (x) = y1 =
(15)
⎩ h 2 (x) = y2 = 1 φr2α + φrβ
2
2
Using the Lie derivatives to express the machine model by the relation input-output
[2, 23].
• Rotor speed y1 =
⎧
⎪
⎪ y1 = L 0f h 1 (x) = h 1 (x) =
⎪
⎨ pM fr Cr
ẏ1 = L f h 1 (x) = φr α I Sβ − φrβ I Sα − − (16)
⎪
⎪ J Lr J J
⎪
⎩
ÿ1 = L 2f h 1 (x) + L g1 L f h 1 (x).u sα + L g2 L f h 1 (x).u sβ
Where
⎡
⎤
Rs 1−σ 1 fr
− + + + I Sβ .φr α − I Sα .φrβ −
pM ⎢ ⎢ σ Ls σ Tr Tr J ⎥ 2
⎥ + f + f Cr
L 2f h 1 (x) = ⎣ ⎦ 2
J Lr pM. 2 + φ 2 − p I φ + I φ J J2
. φrβ α sα rβ sβ r α
σ L s Lr r
PM PM
L g1 L f h 1 (X ) == − φrβ ; L g2 L f h 1 (X ) = − φr α
J σ Ls Lr J σ Ls Lr
2 Rs 1−σ 3M 2M 2 4 2
L 2f h 2 (x) = − M + + Isα φrβ + Isβ φrα + + 2 φrβ + φrα
2
Tr σ Ls σ Tr Tr σ Tr L s L r
2 Tr
2pM 2M 2 2
+ Isα φrβ − Isβ φrα + 2 Isα + Isβ
2
Tr Tr
2M 2M
L g1 L f h 2 (x) = φrβ ; L g2 L f h 2 (x) = φr α
σ L s Tr σ L s Tr
⎧
⎪ z 1 = y1 = h 1 (x) ⎧
⎪
⎪ = ẏ1 = L f h 1 (x)
⎪
⎪ z 2 = L f h 1 (x) ⎪
⎪ ż 1
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ = ÿ1 = L 2f h 1 (x) + L g L f h 1 (x)Vsα + L g L f h 1 (x)Vsβ
⎪ ⎪ ż 2
⎨ z 3 = y2 = h 2 (x)
⎪ ⎪
⎨
z 4 = L f h 2 (x) Derivative ⇒ ż 3 = ẏ2 = L f h 2 (x)
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪
⎪
⎪ ⎪
⎪ ż = ÿ2 = L 2f h 2 (x) + L g L f h 2 (x)Vsα + L g L f h 2 (x)Vsβ
⎪
⎪ ⎪ 4
⎪
⎪
⎪ φ ⎩
⎪
⎩ z 5 = y3 = tan −1 rβ ż 5 = ẏ3 = L f h 3 (x)
φrα
(18)
The determinant of D(x) is non-zero except when the engine is stopped; thus the
matrix D(x) is reversible. The optimal control is given by Eq. (20) [16, 24]:
Vsα −1
−L 2f h 1 (x) + v1
= [D(x)] . (20)
Vsβ −L 2f h 2 (x) + v2
If the determinant of the decoupling matrix is not zero, the non-linear control law
is defined by a relation that connects the new internal inputs (v1 , v2 ) to the physical
inputs (us1 , us2 ) (usα , usβ ) [2, 18, 24]. In order to further a reference trajectory *
and the flux φ*r 2 these variations v1 and v2 are calculated as follows:
⎧
⎪ pM Cr
⎪ ∗ ˙∗ ¨
⎨ v1 = −k11 − − k12 J L φrα Isβ − φrβ Isα − J − + ∗
⎪
r
(21)
⎪
⎪
⎪ 2 2
⎩ v2 = −k21 φrα
2
+ φrβ2
− φr∗2 + k22 M φrα Isβ + φrβ Isα − φrα + φrβ
2
− φr∗2 + φr∗2
Tr
The coefficients k ij are chosen such that the two polynomials s2 + sk 11 + k 12 and
s + sk 21 + k 12 have roots at negative real part.
2
In order to confirm the performances of the proposed control, the numerical simu-
lation was presented on a water pumping system described in section II, composed:
Wind turbine, PMSG, Rectifier, Inverter and a centrifugal pump driven by IM. The
induction motor is controlled by Input-Output Linearization Control (IOLC).the
simulation results are presented by Figs. 2, 3, 4, 5, 6 and 7.
From the simulation results we notice that the Fig. 2 shows the actual and reference
rotor speed. The rotor speed tracks quickly the reference speed.
Control of Wind Water Pumping … 1275
Wr
50 Wref
Wr Wref [Rad/s]
0
-50
-100
1.4
Fig. 3 Rotor flux of IM Phir
Phiref
1.2
0.6
0.4
0.2
0
0 0.5 1 1.5 2 2.5 3 3.5 4
Time [s]
30
Fig. 4 Torque pump and Te
Torque of IM Tpump
20
10
Te Tpump [N.m]
-10
-20
-30
0 0.5 1 1.5 2 2.5 3 3.5 4
Time [S]
40
Fig. 5 Stator current of IM
30
20
Isa [A]
10
-10
-20
0 0.5 1 1.5 2 2.5 3 3.5 4
Time [S]
1276 A. Loubna et al.
1.5
Flow [m3/s]
1
0.5
0
0 0.5 1 1.5 2 2.5 3 3.5 4
Time [s]
speed 30
25
20
Head [m]
15
10
0
0 0.5 1 1.5 2 2.5 3 3.5 4
Time [s]
Figure 3 shows manifests that the quick response of the rotor flux in the transient
state and in steady state the rotor flux is very tracking to the reference flux and is
not affected by resistances variation. We show clearly that the decoupling control
between torque and flux.
Figure 4 shows good torques responses in steady state. It can be observed that the
motor torque is not affected by the rotor speed change and is tracks with very fast
dynamic the load torque of centrifugal pump.
show the wave form that phase stator current is sinusoidal with in acceptable
limits of nominal current of the IM, in transient state the stator current has the peak
current due low values of speed.
See Fig. 6. It is also noted that the flow of the pump are not disturbed by the vari-
ations of the wind speed. The monomeric height is well maintained for the variation
of the speed rotor see Fig. 7.
So from these results we find that the dynamics of flux and the torque is stably
maintained for the different operating regimes, which show that the decoupling is
perfect and assured by this technique. It can be concluded that this technique is
robust and guarantee a good insensitivity to parametric variations of the machine. In
addition, it has the advantage of being simple to implement in practice.
Control of Wind Water Pumping … 1277
5 Conclusion
We have presented in this article, the application of control theory by input output
linearization for a wind turbine water pumping system, shows that the good regulation
of the rotor flux and speed of IM by eliminating system’s nonlinearity and coupling
effect control.
Simulation results have shown the excellent performances of the proposed IOLC
controller. It can be inferred that this system is an ideal solution for water supply in
sparsely populated, isolated and enclosed areas.
References
1. Chikh A, Chandra A (2009) Optimization and control of a photovoltaic powered water pumping
system. In: IEEE Electrical Power & Energy Conference (EPEC), pp 1–6 (2009)
2. Meziane S, Atarsia L, Toufouti. R (2018) A global stability of linearizing control of induction
motor for PV water pumping application. Int J Syst Dyn Appl (IJSDA), 7:31–56
3. Camocardi P, Battaiotto P, Mantz R (2008) Wind generator with double stator induction
machine. Control strategy for a water pumping application.In: 43rd international universities
power engineering conference, pp 1–5 (2008)
4. Rahrah K, Rekioua D (2015) Economic study of a solar/wind water pumping system with
battery storage installed on the site of Bejaia. In: 3rd International renewable and sustainable
energy conference (IRSEC), pp 1–6
5. Bouzeria H, Fetha C, Bahi T, Abadlia I, Layate Z, Lekhchine S (2015) Fuzzy logic space
vector direct torque control of PMSM for photovoltaic water pumping system. Energy Procedia
74:760–771
6. Jaziri S, Jemli K (2013) Optimization of a photovoltaic powered water pumping system. In:
International conference on control, decision and information technologies (CoDIT), pp 422–
428
7. Hammadi S, Hidouri N, Sbita L (2011) A DTC-PMSG-PMSM drive scheme for an isolated
wind turbine water pumping system. Int J Res Rev Electr Comput Eng (IJRRECE) 1:1–6
8. Girma M, Molina M, Assefa A (2015) Feasibility study of a wind powered water pumping
system for rural Ethiopia (2015)
9. Chandel S, Naik MN, Chandel R (2015) Review of solar photovoltaic water pumping system
technology for irrigation and community drinking water supplies. Renew Sustain Energy Rev
49:1084–1099
10. Lebsir A, Bentounsi A, Benbouzid M, Mangel H (2015) Electric generators fitted to wind
turbine systems: an up-to-date comparative study
11. Mahersi E, Kheder A (2016) Adaptive backstepping control applied to wind PMSG system.
In: 7th International Renewable Energy Congress (IREC), pp 1–6
12. Tahour A, Aissaoui A, Abid M, Essounbouli N, Megherbi AC (2010) La commande de la
puissance active et réactive d’une éolienne à génératrice synchrone. In: Revue des Energies
Renouvelables SMEE 2010 Bou Ismail Tipaza, pp 327–335
13. Atarsia L, Oukaci A, Toufouti R, Meziane S (2015) Direct torque control of induction motor
and regulation speed using the adaptive control. In: International conference on automatic
control, telecommunications and signals (ICATS15), pp 1–7
14. Meziane S, Toufouti R, Benalla H (2008) Generalized nonlinear predictive control of induction
motor. Int Rev Autom Control 1:65–71
1278 A. Loubna et al.
15. Toufouti R, Meziane S, Feddaoui O, Saidia M (2016) A robust decoupling control of induc-
tion motors based adaptive rotor flux observer. In: 4th International Conference on Control
Engineering & Information Technology (CEIT 2016), Hammamet, Tunisia, 16–18 December
2016
16. Oukaci A, Toufouti R, Dib D, Atarsia L (2017) Comparison performance between sliding mode
control and nonlinear control, application to induction motor. Electr Eng 99:33–45
17. Tarbouchi M, Le Huy H (1998) Nonlinear control of an induction motor in magnetic saturation.
In: Conference record IEEE industry applications conference. Thirty-third IAS, pp 648–654
18. Pietrzak-David M, De Fornel B, Purwoadi M (1998) Nonlinear control for sensorless induction
motor drives. In: International conference on power electronic drives and energy systems for
industrial growth, Proceedings, pp 300–306
19. Bendaha Y, Mazari B (2007) Commande Adaptative Linéarisanted’un Moteur Asynchrone
20. Blaschke F (1972) The principle of field orientation as applied to the new transvector closed-
loop system for rotating-field machines. Siemens Rev 34:217–220
21. Casadei D, Profumo F, Serra G, Tani A (2002) FOC and DTC: two viable schemes for induction
motors torque control. IEEE Trans Power Electron 17:779–787
22. Zhu J, Xu B, Wang X, Feng H, Xu X (2013) The research of sensorless vector control for
permanent magnet linear synchronous motor. J Comput 8:1184–1191
23. Enev S (2007) Input-output decoupling control of induction motors with rotor resistance and
load torque identification. In: Mediterranean conference on control & automation, pp 1–5
24. Marino R, Peresada S, Tomei P (1999) Global adaptive output feedback control of induction
motors with uncertain rotor resistance. IEEE Trans Autom Control 44:967–983
25. Harrouz A, Dahbi A, Harrouz O, Benatiallah A (2014) Control of wind turbine based of PMSG
connected to water pumping system in South of Algeria. In: 3rd international symposium on
environmental friendly energies and applications, pp 1–4 (2014)
26. Dahbi A, Harrouz A. MPPT control of wind turbine for water pumping system. Relation 7:0.1
27. Charles CR, Vinod V, Jacob A (2015) Field oriented control of Doubly Fed Induction Generator
in wind power system. In: IEEE international conference on computational intelligence and
computing research (ICCIC), pp 1–5
28. Pantea A, Sivert A, Yazidi A, Betin F, Carriere S, Capolino G-A (2016) Efficient field oriented
control with power losses optimisation of a six-phase induction generator for wind turbines.
In: IECON 42nd annual conference of the IEEE industrial electronics society, pp 1912–1917
29. Rebei N, Benghanem B, Hmidet A, Hasnaoui O (2013) Study of photovoltaic water pumping
system using scalar-DVC based control. In: International conference on electrical engineering
and software applications, pp 1–8
30. Guettaf A, Bettka A, Bennis O, Benchabane F, Khaled Y, Ali A (2011) Direct field oriented
control of induction motor fed by wind turbine generator under saturation effect
31. Toual B, Mokrani L, Kouzou A, Machmoum M (2018) Control and management of a solar-wind
hybrid system for power quality improvement. J Eng Sci Technol 13:1420–1439
32. Guettaf A (2013) Optimisation floue-génétique d’un système de pompage éolien.Université
Mohamed Khider–Biskra
33. Zhang J, Qi Z, Wen J (2017) Control of water pumping driven by wind turbine based on Quasi
Z source inverter. In: Chinese automation congress (CAC), pp 7074–7077
34. Atasrsia L, Toufouti R, Meziane S (2020) Standalone photovoltaic array fed induction motor
driven water pumping system. Int J Electr Comput Eng (2088–8708) 10
35. Belarbi M (2006) Modeling and simulation of PV water pumping system. Master thesis Oran
university
36. Karampuri R, Jain S, Somasekhar V (2014) A single-stage solar PV power fed open-end
winding induction motor pump drive with MPPT. In: IEEE international conference on power
electronics, drives and energy systems (PEDES), pp 1–6
Improved Hybrid Control Strategy
of the Doubly-Fed Induction Generator
Under a Real Wind Profile
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1279
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_117
1280 M. Yessef et al.
1 Introduction
All over the planet, the demand of electrical energy is increased more and more.
However, the major energy resources of electricity generation are natural gas, fossil
fuels, petroleum [1]. Nowadays, the excessive consumption of these fossil sources
causes a serious environmental and economic problems; namely, on the environ-
mental side, planet temperature increases and global warming, water pollution, the
emission of CO2 and other toxic gases, and acid rain. For the economic impact, it
consists of the increase in the price of petrol, oil and gas, which influences the elec-
tricity tariff and, subsequently, on all economic, agricultural and service activities.
Generally, this huge use has a dangerous effects on humans and animals.
Therefore, the renewable energy resources are revealed to be the serious alter-
natives to the polluting energies [2]. There are many kinds of renewable energy
resources, that are exploited all over the world (e.g. solar, geothermal, biomass,
hydropower and wind) [3]. In the last decades, the installation of wind farms has
been developed and received a great support from the most governments, thanks to
their profitability and cleanliness. Then, to face these urgent energy challenges, the
wind energy becomes usual more and more in many countries with global power
reaching over 600 GW and it target to be 1600 GW in 2030 [Word Wind Energy
Association]. So, the optimization of the operation of the wind energy conversion
system (WECS) has become essential, given their large share of the global energy
production.
Thereby, due to their several advantages, the DFIG (Doubly-Fed Induction Gener-
ator) technology with variable speed is presented to be the important way to develop
turbine generators within other competing machines. The DFIG is characterized by
its ability to work over a wide range of wind speeds and it performs an optimal
conversion of aerodynamic power at a low cost [4]. The Kinetic power of the wind
is transformed to mechanic power by the turbine and then delivered to the electrical
machine by the gearbox in the transmission machine. The DFIG stator is directly
linked to the electrical grid, at a time, the rotor is connected by an electronic power
converter. Since the stator of the DFIG is connected directly to the electrical grid,
when high peaks of stator currents occur at the stator windings, because of the estab-
lishment of a high voltage drop, overcurrents occur at the rotor windings and then on
the Rotor Side Converter (RSC), due to the coupling between the stator and rotor [5,
6]. For this reason the DC bus voltage increases and the torque fluctuations would
take place. Therefore, in the absence of the protection circuit, the high transient
currents will damage the RSC and the DFIG [5, 7].
Parallel to the wind farms development, several algorithms and structures of
control are then proposed to confident the stability, the performance and the robust-
ness of aero-generator of WECS such as the PI (Proportional Integral) controllers
and the adaptive PI [9, 10]. Other works focused on the other command types such
as the scalar, the adaptive fuzzy logic controller, the direct power control (DPC) [5],
Improved Hybrid Control Strategy ... 1281
the sliding mode control (SMC) [11, 12], the predictive control and the lineariza-
tion control. Unfortunately, the simulations with real wind and the testing of several
commands on the same system in the same environment are lacking, in this research
fields.
In this work, an improved command has been developed, which combines between
the Direct Power Control and the Backstepping command. The utilization of real
wind conditions from a wind farm located in south of Morocco is another innovative
aspect meted in this research paper.
This research paper is presented as follow; firstly, after the introduction, the turbine
is modeled and the dynamical model of the DFIG is presented, then the general
structure of the hybrid DPC-BS command is approached. The last section for this
paper will be reserved for obtained simulation results of performance and THD tests.
All results are reported and have been analyzed to prove the validity, of the proposed
command, to be implemented in a real scenario.
2 Methods
The Fig. 1 shows the structure of the studied WECS. This system is composed of a
wind turbine driving a DFIG connected to a three-phase electrical grid. This model
is characterized by the fact that the stator is directly coupled to the grid, whereas,
in order to establish a power exchange, between the machine and the grid, in sub-
synchronous speed, two converters are installed between the electrical grid and the
rotor. As the active and reactive powers of DFIG are controlled through the RSC, the
GSC operates as a rectifier [5, 7].
The wind energy is captured by the wind turbine and converted into a torque or speed.
This process is modeled by the following equations [7]:
⎧
⎪
⎪ PRotor = Pwind .C p (λ, β) ⎧
⎪
⎨ ⎨ PRotor = 0.5ρ.A.C p (λ, β).v
3
Pwind = 0.5ρ Av 3
⇒ 1 (1)
⎪
⎪ PRotor ⎩ TRotor = .ρ.S.C p (λ, β).v 3
⎪
⎩ TRotor = 2t
t
⎧
⎨ C (λ, β) = k .( k2 − k .β − k ).e( −kA5 ) + k .λ
p 1 3 4 6
where: A
⎩
k1 = 0.5872, k2 = 116, k3 = 0.4, k4 = 5, k5 = 21, k6 = 0.0085
At the rotor of the turbine, the expression of the mechanical torque is [13]:
⎧
⎨T dmec
mec =J dmec
dt ⇒J = Tg − Tem − f mec (2)
⎩ dt
Tmec = Tg − Tem − f mec
g = G.t
where the gearbox equations are given by:
TRotor = G.Tg
In the WECS, the mechanical power is transferred to electrical power by the DFIG.
In the (d,q) reference frame, the electrical equations of the stator and the rotor which
model the DFIG are given by [7]:
⎧ ⎧
⎪ dϕsd ⎪ dϕr d
⎪
⎪ V = R .I + − ω .ϕ ⎪
⎪ Vr d = Rr .Ir d + − ωr .ϕrq
⎪
⎪
sd s sd
dt
s sq
⎪
⎪ dt
⎪
⎪ ⎪
⎪
⎪
⎪ dϕsq ⎪
⎪ dϕrq
⎪
⎨ Vsq = Rs .Isq + + ωs .ϕsd ⎪
⎨ Vrq = Rr .Irq + + ωr .ϕr d
dt dt
and (3)
⎪
⎪ 1 Msr ⎪
⎪ 1 Msr
⎪
⎪ Isd = .ϕsd − .ϕr d ⎪
⎪ Ir d = .ϕr d − .ϕsd
⎪
⎪ σ.L σ.L ⎪
⎪ σ.L σ.L r .L s
⎪
⎪
s r
⎪
⎪
r
⎪
⎪ 1 Msr ⎪
⎪ 1 Msr
⎩ Isd = .ϕsd − .ϕr d ⎩ Irq = .ϕrq − .ϕsq
σ.L s σ.L s .L r σ.L r σ.L r .L s
According to the latest equations, the rotor voltages will be the instructions to be
given to the RSC to control the active and reactive stator powers. Thus the reactive
stator power Qs will be controlled by the rotor voltage Vrd , and the active stator power
Ps will be controlled by the rotor voltage Vrq . On the other hand, the powers errors
eQ and eP , used in the proposed DPC-BS algorithm, are given by [5]:
e Q = Q ∗s − Q s
(10)
e P = Ps∗ − Ps
1 2
v= e Q + e2P (12)
2
Improved Hybrid Control Strategy ... 1285
The Backstepping controller intervenes when the voltage dips occur on the stator
supply terminal, in order to calculate the stator power references to control the RSC
and push the DFIG to operate correctly and satisfactory, by imposing the return of
the voltage at its nominal value (Fig. 2).
[4]. The parameters of the used machine are described in the Tables 1 and 2 (see
appendix).
This test allows to evaluate the performance, the reliability of the WECS and the
decoupling of the powers generated by the DFIG during the instantaneous variation
of the power set-points (Fig. 4, 5, 6, 7).
The analysis of the simulation results confirm, that the active power follows
perfectly its reference value, which is deduced by the MPPT strategy, with a dynamic
error of εd = 1.05%. It can be also seen, that the reactive stator power is kept null in
order to have a unit power factor obtained after the end of its transient regime. More-
over the stator currents varies according to the wind variations, which confirms its
proportional relationship with the power. Despite the wind fluctuations, this current
zoom
15
10
0 5 10 15 20
10.5
10
Zoom wind speed (m/s)
9.5
8.5
7.5
0 5 10 15 20 25
time (s)
1500 -500
Ps Ps*
0
zoom
Actice power (w)
-500 -1000
-1000
Zoom
-1500
-1500
-2000 -2000
-2500
-3000 -2500
0 5 10 15 20 25 14.5 15 15.5
time (s) time (s)
10
5
0
-5
-10
-15
0 5 10 15 20 25
time (s)
Improved Hybrid Control Strategy ... 1287
-40
zoom(1)
-80
0 5 10 15 20 25
time (s)
10
8 ira irb irc
Rotor currents (A)
6
4
2
0
-2
-4
-6
-8 zoom(1)
-10
0 5 10 15 20 25
time (s)
remains sinusoidal with a frequency of 50 Hz, which is the frequency of the electrical
grid. The rotor currents remain sinusoidal during the DFIG operation, its frequency
varies proportionally to the mechanical speed. These currents reveal fewer ripples
compared to other control techniques existing in literature.
In order to verify the effect of the hybrid command DPC-BS on the quality of the
supplied power by the machine, a spectral analysis of the stator and rotor currents
was carried out. Note that this measurement was carried out for 3 operation cycles
with a harmonic order of 70. The Fig. 8 shows respectively the currents of the stator
phase “isA ” and the rotor phase “ira ” supplied by the DFIG, as well as their harmonic
spectra.
These figures confirmed the influence of the control on the quality of the supplied
energy. The rate of harmonic distortion is equal to 0.33% at the level of the rotor
current, and 0.94% at the level of the stator current. From the depicted results in
these simulations, it can be noted that the harmonic distortion rate THD has a good
quality for the both injected currents into the electrical grid and remains below the
limit of 2% which imposed by the IEEE-519 standard [13].
This THD value (0.33%) obtained, in this work, by the proposed DPC-BS hybrid
control is the lowest, compared to the value recently reached by the researchers [8]
for WECS based on DFIG which is 1.67%.
4 Conclusion
In this work, we have presented a contribution for the control of the wind energy
conversion system. Therefore we applied the improved hybrid DPC-BS command
to improve the performance of the doubly-fed induction generator. This approach
is based on the combination of the Backstepping controller and the Direct Power
Control. The simulation results analysis have shown that the hybrid DPC-BS
command gives highest dynamic performance in the reference tracking, in a test
scenario with a real wind profile, and it also ensures an optimal energy conversion.
Decidedly, this command provides, a concrete solution to the problems of robustness
and pursuit. Moreover, another essential point of this work is that, it confirmed the
injecting possibility of the supplied energy, by the studied WECS, into the electrical
grid.
For the future works, a hardware-in-the-loop simulation on dSPACE will be
realized to verify the proposed approach in the real time by experiments.
Appendix
Table 1 (continued)
Name Values
Rotot resistance Rr = 0.816
Stator inductance coefficient Ls = 0.071H
Rotor inductance coefficient Lr = 0,070H
Mutual inductance coefficient M = 0,069H
Friction coefficient f = 6,73.10−3 kg.m2
Moment of inertia J = 0,3125 N.s/rad
References
Abstract In this article we have studied the control of a high performance electric
machine, more precisely a permanent magnet synchronous generator using the Flux
Orientation Control (FOC), this allowing us to independently control the torque and
the flux of our machine in the same way as a DC machine with separate excita-
tion, where the inducing current controls the magnetic flux, and the induced current
controls the electromagnetic torque.
We transformed the stator instantaneous currents to two current components, one
which controls the flux (along the d axis), and the other controls the torque (along
the q axis).
The goal of our article is based on the DFOC (direct) technique, which consists
of neglecting the coupling terms, and setting up a proportional integral regulator (PI)
in order to follow the parametric variations of our machine using two types of wind
(variable and step) and independently control the active and reactive power.
1 Introduction
In the last few years, the use of synchronous machines in industry has grown sharply
because of their similar characteristics with DC machines [1, 2].
Two types of synchronous machines are present in the market, synchronous
machines with permanent magnets and winding rotor machines, those with perma-
nent magnets have superior performance compared to the other with excellent
efficiency, high mass power and reduced maintenance [3, 4].
Several control strategies are present in the literature and are reserved to
synchronous generator with permanent magnet, we can quotes for example:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1291
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_118
1292 N. Z. Laabidine et al.
direct torque control (DTC), vector oriented flow control (FOC) or the non-linear
Back-stepping control.
We present in this article the modeling of the synchronous permanent magnet
machine using vector-oriented flow control [5] (FOC) applied to the static converter
on the machine side [6], which allows us to control the electromagnetic torque,
magnetic flux, active and reactive powers of a wind generator in order to obtain a
simple and reliable model for numerical simulation.
The GSAP has a coiled stator but the rotor is replaced by magnets, which eliminates
rhetoric losses and is considered a power source that can provide excitation current.
With:
• R s :The resistance of a statoric phase ()
• V sa : Statoric phase voltage (V)
• i sa : Statoric current intensity (A)
• sa : Total flow through the stator coils (Wb)
PMSG stator voltage equations in a system of axes (d-q)
With:
• ωe : The electric pulsation of voltages (rad.s−1 )
• sd : Direct component of the statoric flow (Wb)
• sq : Quadratic component of the statoric flow (Wb)
Flow-Oriented Control Design of Wind Power Generation System ... 1293
sd = L d · i sd + f (4)
sq = L q · i sq (5)
With: L d L q are the direct and quadratic components of the statoric inductance
(H).
Then [8]:
d L d · i sd + f
Vsd = −Rs i sd − − ωe L q · i sq (6)
dt
d L q · i sq
Vsq = −Rs i sq − + ωe L d · i sd + f (7)
dt
Ignore the stator transients and obtain:
d(i sd )
Vsd = −Rs i sd − L d . − ωe (L q · i sq ) (8)
dt
d i sq
Vsq = −Rs i sq − Lq · + ωe L d · i sd + f (9)
dt
From (8) and (9) we can deduce the equations of statoric currents
d(i sd ) −1
= (Vsd + Rs i sd + ωe (L q · i sq )) (10)
dt Ld
d(i sq ) −1
= (Vsq + Rs i sq − ωe L d · i sd + f ) (11)
dt Lq
−3
Tem = .p. sd .isq − sq· isd (12)
2
−3
Tem = .p. (Ld − Lq )isd isq + f· isq (13)
2
−3
Tem−r e f = 2
.p.f .i sq (For a machine with plain poles(L d = L q )).
1294 N. Z. Laabidine et al.
Vsq = −(R s + S.L q ).I sq + ωe L d .Isd + f (15)
−1
Isd = (Vsd + ωe (L q · i sq )) (16)
(R s + S.L q )
−1
Isq = (Vsq − ωe L d .i sd + f ) (17)
(R s + S.L q )
3
Pgen = Tem . = . Vsd .i sd + Vsq .i sq (18)
2
3
Q gen = . Vsd .i sd − Vsq .i sq (19)
2
Equivalent circuit (d-q) for a PMSG (Fig. 1)
Vector control was first proposed by “Blaschke” in 1972 [1], after it was developed
by “Hasse” in 1979 [2], also called flow-oriented control (FOC), is a method in which
the three-phase statoric currents of an AC electric motor are transformed into two
orthogonal components which can be considered as vectors, The first produces the
electromagnetic torque and the second produces the magnetic flux. This allows us to
operate similar to a DC motor [10].
An Integral Proportional Controller (PI) was used to maintain the current at its
requested value (Id) (Fig. 2).
Flow-Oriented Control Design of Wind Power Generation System ... 1295
The kinetic energy of the wind is transformed into mechanical energy by the rotor
of the turbine, and then converted into electrical energy by a generator; its formula
is given by [11]:
1
Ec = .m.v 2 (20)
2
The kinetic power of the wind received by the turbine through its blades is given
by:
1
Pwind = .ρ.π.R 2 .v 3 (21)
2
The aerodynamic power Ptur appearing at the turbine rotor is written:
1
Ptur = .C p (λ, β).ρ.π.R 2 .v 3 (22)
2
The expression of the aerodynamic torque Cp of the turbine is given by the
following formula:
1
Cp = .C p (λ, β).ρ.π.R 2 .v 3 (23)
2.t
Ptur = ω.T
tur ; λ = V
Rω
2
Ttur = 21 ρπ R 3 V C P (λ, β)
1296 N. Z. Laabidine et al.
5 MLI Model
In order to have a constant frequency and amplitude voltage on the machine side and
to optimize the power taken from our wind energy conversion system, we will use
MLI power converters.
⎧
⎨ +1, S = −I
S= , S = a, b, c (24)
⎩
−1, S = +I
Park Transformation
Park’s transformation is also called direct-quadrature-zero (dq0) allows to transform
the stator or rotor windings a, b, c into two windings d and q, We denote by “d” the
direct axis and “q” the quadrature axis.
6 PI Regulator
In order to regulate the stator current values i sd eti sq , we are going to use a PI regulator,
which also aims to provide better performance and robustness to our wind power
k .s+K
system. The equation of the PI regulator used to control the GSAP: p s i
With:
• k p : proportional gain of the corrector PI
• K i : proportional gain of the corrector P
d(i sd )
Vsd = −Rs Isd − L d . − ωe (L q · i sq )
dt
d(i sq )
Vsq = −Rs i sq − L q . + ωe (L d .i sd + f )
dt
We put Aq = ωe (L q · i sq ); Ad = ωe (L d .i sd + f )
1298 N. Z. Laabidine et al.
1 1 1
FT(s) = = . ;
Rs + L s .s Rs 1 + Te .s
with Te = L s Rs (Electric time constant)
L d = L q (Smooth pole machine stator inductance) that implies that k p = ki
(Fig. 3 and 4).
We can simplify the control loop to be similar to that of a first order system (Fig. 5).
The open-loop transfer function is given by:
k p .s + K i 1
FTBO(s) = (32)
s Rs + L s .s
1 1
FTBF(s) = = (33)
Rs +k p
.s
Ls 2
ki
+ ki
s+1 1
ωn
.s 2 + ω2ζn s + 1
ki = ωn 2 .L s ; k p = 2.ζ.ωn .L s − Rs
8 Conclusion
References
Abstract Careful planning and forecasting of energy consumption not only influ-
ences a nation’s environmental and energy sustainability, as well as giving a useful
basis for policy makers to make decisions. This paper presents the results of an appro-
priate deep learning model forecasting for consumption using echo state network
(ESN). ESN is a new paradigm that offers an intuitive methodology using for time
series prediction. Basically, it is a recurrent neural network (RNN) with a vaguely
connected hidden layer, known as a reservoir, which functions in a strange way in
the existence of time-series patterns. In this contest, three types of recurrent neural
network used for comparison with ESN, aiming to evaluate the accuracy and perfor-
mance of the model. To train and test the proposed model, we used the historical data
of multivariate household consumption. ESN showed an improvement of 0.057% and
0.095% in terms of mean square error (MSE) and mean absolute error (MAE), and it is
trained faster than other models, in very short-term energy consumption forecasting.
1 Introduction
The role of energy in human life is absolutely essential. In one way, the evolution of
human society due to the rise of energy efficiency and the exploitation of new artificial
intelligence techniques. In parallel with economic and population dynamics, global
consumption of energy continues to increase, causing several challenges. There is
enormous pressure on the energy sectors worldwide to ensure a stable and reliable
energy supply. Over the years, the consumption of energy has grown remarkably.
Based on the statistics published in 2019 by British Petroleum, there will be a 2.8%
increase in primary energy consumption in 2018, which is higher than in the previous
year 2017 [1]. Due to the rapid growth of intelligent measuring systems and the
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1305
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_119
1306 W. Bendali et al.
“Green Button” project [2], electricity consumption power for household users is now
measured, collected and presented. The use of Demand Response (DR) and Home
Area Networks (HAN) has put a new emphasis on individual households [3]. While
residential households contribute to DR programs as a willing-seller, it is difficult
to plan and estimate energy for these individual households prior to the program
without an energy forecast at the individual household level. Certain utilities publish
utilization or billing forecasts for residential customers to provide an innovation in
service, assisting households to plan their consumption and reduce their energy bills.
As a result, there is a growing recognition of the potential for forecasting household
energy consumption by governments and research institutes. Short- and very short-
term energy forecasting methods are very useful for managing residential energy
demand, energy efficiency, and market design of electricity prices, in addition they
are very powerful for planning in maintenance of large-scale, complex smart grids
[4–6]. Furthermore, it offers the smart grid reliability, security and protection to help
meet the increasing energy demand of residential homes.
ESN represents a family of Reservoir computing (RC) models where the recurrent
part is created randomly, and then kept fixed. Even with this strong simplification, the
recurring part of the model (the reservoir) provides a rich set of dynamic character-
istics that allow to solve a wide variety of problems. In deep learning, the echo state
network (ESN) further belongs to the general class of the reservoir computing. The
ESN has powerful non-linear time series modelling capabilities ESN has a dynamic
reservoir with a number of connected units, which makes the calculation of the data
very uncomplicated [7]. Among the specific features of the reservoir computing
methods, they have the ability to train only the readout weights by using the method
of linear regression throughout the training procedure. By condensing this character-
istic, ESN can achieve a high velocity of convergence towards an excellent overall
solution. This study aims to propose echo state network model with multivariable
input for energy consumption prediction.
The rest of this study is organized as follows. Section 2 introduces consump-
tion data description and describes preprocessing of data. Section 3 represents the
basic ESN architecture, his parameters and training process. Section 4, experimental
application used, and presentation of results. Finally, conclusion.
2 Data Description
Generally, four time horizons can be distinguished, that are used to predict time series:
very short, short, medium and long term [8]. Very Short-term prediction concentrates
on minute time ranges, contributing to assure an uninterrupted power flow in the near
future and supporting electricity trade and price management [8, 9]. Medium and
long-term prediction concentrate over a duration of a few days to several months and
Households Energy Consumption Forecasting with Echo State Network 1307
assist in the installation and maintenance of production facilities [8, 10], transmission
and distribution system development [11], and energy plan formulation [12]. In this
work, very short term data used for train forecasting model, which is a multivariate
time series of household electricity consumption. The data used are collected from
a house located in Sceaux (7 km of Paris, France), during one year (2009), with
observations of every minute [13]. These data made up of 7 features:
Pa :: Global active household power consumption (KW).
Pr : Overall reactive power required by the household (KW).
V : Means voltage (volts).
Ig : Average current intensity (amps).
E a1 : Sub-measure of active energy for the kitchen (Wh).
E a2 : Sub-measure of active energy for laundry (Wh).
E a3 : Sub-measure of active energy for climate control systems (Wh).
The data set contains the active power, voltage and intensity current averages
and certain distribution of active energy in the household like kitchen, washroom
and air-conditioning. In order to obtain greater accuracy, we also add the rest of the
sub-measures of the active energy E a4 , subtracting the three sub-measures from the
overall active energy, by this following equation (Eq. (1)). All features are shown in
Fig.1:
E a4 = (P a × 1000/60) − (E a1 + E a2 + E a3 ) (1)
In this work, we use data of 1 year (2009), which contain timescale for each minute.
This data are separated like this: 473040 sample used for model train, 52560 samples
used for the test. Global active power data have strong nonlinear properties, this can
cause negative impacts on the prediction model. Before the model is fed by train
data, we pre-process input values, for raison to minimize the effect of the wide values
scale. The pre-processing process that we choose, namely the linear normalization
that changes the data in the interval between 0 and 1. Post-processing is necessary
upstream or downstream of the analysis for the performance of the forecast model.
In this case post-processing of the normalized values must be anti-normalizing to
extract the expected actual global power active and analyze the model performance.
1308 W. Bendali et al.
3.1 Architecture
The architecture of ESN is like any neural network, has input, hidden, and output
layers, which are also having a certain number of units (neurons). However, as shown
in Fig. 2, ESN has a reservoir instead of simple hidden layer, made up of some number
of recurrent or simple units. These units are interconnected in order to achieve high
treatment capability and, consequently, obtain good forecasts [7]. Although the ESN
has been applied in many areas since its introduction in 2001 [14], this includes time
series prediction and data classification [7, 15]. Generally, ESN consists of k input
neurons that represent features or variable numbers, n units in the hidden reservoir,
which are responsible for creating internal information, and o units in the final layer,
which gives forecast time series values. This study deals with prediction problems
in a single step so we are used o = 1, in time step j, assume that the input array of
the input state is given as Eq.(2) The array of the reservoir state is Eq. (3), and the
array of the output state is Eq. (4):
In ESN, four weight matrices exist, win represent the connection between input
and reservoir state, w represent the weight between internal units of reservoir layer,
wback gives the connection between neurons of output layer and reservoir layer, and
wout as a matrix weight that link the internal units and output units. The activation
functions of the reservoir and output layers is F and G respectively. These functions
can represent a linear or nonlinear function depending on the nature of data.
1310 W. Bendali et al.
The weight matrix wout , which is from the reservoir to the output layer, can be
updated throughout the training process, whereas win , w and wback are randomly
chosen then stay unchanged [16, 17].
ESN has three critical reservoir Hyperparametres, which have a major effect on the
efficiency of the system and should be set at appropriate values.
n: The number of units in hidden reservoir, that can effect in terms of model
quality and accuracy, due to his role in dealing with large data size in the training
phase. Generally, the choose of number of units according on the nature and size of
data using for training. In this study, we use n = 60.
α: The connectivity rate, which allow to choose the linking degree between units
of hidden reservoir layers. That can aide to control the memory chance and deal with
computing complexity [16]. In general, with reference to review of experience, α is
chosen between 1%–5% [18].
ρ: The spectral radius, has a very important role in ESN model, his value represents
the max of the absolute proper value of the weight matrix w. In other to make ESN
has dependency only with input and output sample of training we concede that the
range of ρ is between 0 and 1. It is unaffected by its first state after a few epochs if
the entries are long enough. [19, 20].
In order to achieve accurate output compared with target values, which are collected
from data training, we have to train models by using N time samples of input and
target data. There are three steeps using for training models: Firstly, it must initialize
the ESN parameter values, by choosing a pertinent number of reservoir units n, select
the best values of connectivity rate α and spectral radius ρ, and initialize randomly
weight values win ,wback , and w. After that, it is important to update the reservoir
states using the actual input, the previous values of reservoir and output states as
mentioned in Eq. (5). Finally, Computing wout by obtaining M the vector of hidden
reservoir values, which has (N − I0 + 1) × n sample, and T the target values, which
has (N − I0 + 1) × o as follows in Eq. (7). With I0 is the washout time step.
T
wout = (M −1 T ) (7)
ESN has a special training algorithm, which can make it able to deal simply and
quickly with data like consumption power of households [14, 21].
Households Energy Consumption Forecasting with Echo State Network 1311
In this work, we create our model with Python environment by using KERAS as a
deep learning library, which supports CPU and GPU [22]. We choose RNN and their
extension Long Short Term Memory (LSTM), and Gated Recurrent Unit (GRU) as
deep learning models to compare them with ESN. These models used input data that
contains eight variables represent eight units, total active power(GAP), total reactive
power, voltage, total intensity, and four Sub-measuring data of the household’s energy
(Fig. 1). The row of this data contains 525600 samples of minute, these data are used
as inputs of models training and testing. Data are split into 90% for training and 10%
for testing. The output of the models has one unit, which represent, predict global
active power consumed. For dealing with the effect of dimensionality problem of data,
we normalize data values between 1 and 0 by using normalization as a preprocessing
method. After obtaining predict values, we used anti-normalization for obtaining
the forecast GAP. The hyperparameters of ESN that we choose the number of units
n = 50, connectivity rate α = 0.02, and spectral radius ρ = 0.5. We choose these
parameter values based on Trial and Error method, aiming to achieve the smaller
error by testing various parameter values. This method is an efficient and standard
method of parameterization. Activation functions using in ESN are F = Tanh and
G = identity. Concerning the standard RNN and its extensions (LSTM, GRU), has
two hidden layer’s neurons,each one of theme has 60 neurons or units with Tanh
activation function for two hidden layers and Relu for the output layer. The error
metrics compared in this paper was mean square error (MSE) (Eq.(8)), and mean
absolute error (MAE) (Eq.(9)), as following equations:
1
N
MSE = (P f or ecast (t) − Ptr ue (t))2 (8)
N i=1
1
N
M AE = P f or ecast (t) − Ptr ue (t) (9)
N i=1
With N is the number of samples, P f or ecast (t) is the forecasting of global active
power at time t, and Ptr ue (t) is the true global active power at time t.
In this work, we compared the accuracy of the proposed deep learning method
ESN with simple RNN, LSTM and GRU. As we can see in Figs. 3, 4, 5, and 6, ESN
model has more accuracy than standard RNN and his extension in consuming power.
In Table 2, comparisons based on MAE and MSE of models. As shown in this table,
the LSTM and GRU models have better performance than Simple RNN method,
that is because LSTM and GRU methods characteristics of carrying initially learnt
important information over a long distance [23, 24]. However, ESN still much better
compared with them, because of his ability to deal with nonlinearity, and his special
training algorithm. In terms of training speed, we can see from Table 1, ESN is faster
1312 W. Bendali et al.
Fig. 3 True values vs. ESN forecast values of global active power consumed in 1 day
Fig. 4 True values vs. GRU forecast values of global active power consumed in 1 day
Fig. 5 True values vs. simple RNN forecast values of global active power consumed in 1 day
than RNN, which is considerably faster than LSTM, and GRU, that is because in
ESN we only train the output weight wout and it can speed up the training of neural
network.
Households Energy Consumption Forecasting with Echo State Network 1313
Fig. 6 True values vs. LSTM forecast values of global active power consumed in 1 day
5 Conclusion
References
21. Chen Q, Shi L, Na J et al (2018) Adaptive echo state network control for a class of pure-feedback
systems with input and output constraints. Neurocomputing 275:1370–1382
22. Dahl GE, Sainath TN, Hinton GE (2013) Improving deep neural networks for LVCSR using
rectified linear units and dropout. In: 2013 IEEE international conference on acoustics, speech
and signal processing, vancouver, BC, p 8609−8613. https://ieeexplore.ieee.org/stamp/stamp.
jsp?tp=&arnumber=6639346&isnumbe=6637585
23. Ruiz CW, Perapoch J, Castillo F, Salcedo S, Gratacós E (2006) Bessons monocoriònics afectes
de transfusió fetofetal: conseqüències a curt i llarg termini. Pediatria Catalana 66(2):53–61
24. Cho K, van Merrienboer B, Bahdanau D, Bengio Y (2015) On the properties of neural machine
translation: encoder–decoder approaches. pp 103–111. https://doi.org/10.3115/v1/w14-4012
Management of Multi-agents in a Smart
Grid Network with the Python Using
the Contract Net Protocol
Abstract Smart grids are one of the best solutions for the integration of different
distributed energy sources (DES), load, and storage elements. However, optimizing
and managing these systems is a big challenge because it is only done in a distributed
way. One of the best methods of managing a smart grid is the use of artificial intelli-
gence (AI), especially multi-agent systems (MAS) approaches. This work constitutes
a demonstration of one of those methods which is the contract net protocol (CNP).
We will use the extended version of the contract net protocol(ECNP) to optimize
the movement of MASs inside a smart grid network. Our example of a smart grid
will be a grid constituted from nodes, and each node is a location of an element of
the smart grid which could be a wind turbine, a customer…, in other words, a task
or/and an agent. In the scenario each task it’s given a target which is also a node
inside the grid, and our main objective is that the agent or agents could contribute
to each task, depending on their priority order and may allow it to reach its target,
which is another node in the network far from its initial location, using the shortest
way as possible.
1 Introduction
The great and fast development of renewable energy sources (RESs) [1, 2] encouraged
researchers to begin forming different size units provides as microgrids [3]. They
were presented as a solution for the reliable integration of distributed energy resources
[4, 5], and resilience-related problems [6, 7] as shown in Fig. 1.
Developing such outstanding units of energy has so many economic and environ-
mental benefits [8, 9] as reliability because they improve consumers’ supply [10, 11],
their average interruption frequency, duration between consumers and system, and
resilience which is the capability of the system to minimize possible power outage
and returning to stable status [12].
With the start-up of microgrids, different types of control architectures have been
defended, starting with the very first ones as centralized and decentralized procedures.
That kind of approach depends on how many components the microgrid is constituted
of and its dimension, so their complexity increase despite it. Each architecture has
a particular number of levels, and every level supports unique tasks, as maintaining
stability or load sharing. However, the major challenge with microgrids is to provide
efficient management techniques with a high level of optimization results for system
efficiency, minimizing the use and processing of data, and, of course, independence in
management [13, 14]. Yet, those architectures didn’t bring much satisfaction to their
inventors, due to their high cost and management’s difficulties, the reason behind
swishing into decentralized architectures.
On account of the arguments mentioned above, added to the important number
of DERs, loads, and storing devices implemented on MGS [12, 15, 16], shifting into
distributed control methods has been motivated [17], as agent technology has given
their many advantages like using local data exchanges between neighbors, especially
the multi-agents systems (MASs) approaches [18–20], to optimize the operation of
local power production, and improve performances of microgrid using decentralized
control [11].
MASs have been defined in many ways in literature, as «a comprises of two or
more independent agents with some information to achieve a set target» [21], or
autonomous decision-making entities [22]. They are known for their capability of
reaction in the environment, intercommunication inside the network, and high level
of autonomy.
Yet, to reach those advantages and gain from MASs qualities, we should overcome
technical challenges and problems, which are quite numerous [21, 22], starting with
MASs platform choices, communication languages, architecture, and design while
respecting the Foundation for Intelligent Physical Agents standards [23] shown in
Fig. 2. In other words, insuring compatibility, robustness, flexibility, extensibility
plus the capability of interacting with each other irrespective of the platform they
run on [22]. One of those techniques will be presented below.
This article is organized as follows: in the second section we will present a defi-
nition of the contract net protocol, the technique used in our work to optimize the
movement of the agents, after that, we will explicitly explain our work with different
scenarios, tables, and simulation showing the movement inside the smart grid, at last
in the third section we will end up with a conclusion and a discussion.
Before explaining our first scenario, we will clarify our program configuration and
the implementation structure:
Program Configuration
Our configuration is established of six major scripts:
The inputs of our system as shown in the scripts are the tasks and the outputs are
the bides provides by the agents at each scenario.
About Scenario_1_Part_AG:
The first scenario is developed to test our version of ECNP for tasks and agent’s
movement inside a smart grid network. We choose to use one agent and three tasks.
Every task comes at a specific time identification(id), as known that our agent has
a special order of priority, we spread tasks in the opposite order of that priority, the
main goal of that is to show the capability of the agent to bid and contract for a task
no matter its importance for him (Table 1).
1) Task used in the first scenario:
Management of Multi-agents in a Smart Grid Network ... 1321
As shown in the table above, every target is introduced in a special time id, and each
one has its aim that it should reach to complete its role inside the smart grid network
(Table 2).
2) The agent used in the first scenario:
The second table presents our agent’s attributes and characteristics. Depending on
those qualities the agent can bid and contract for every task and offer the best path
between the task first position and its aim position.
3) Smart grid network used in the first scenario studied:
The grid extracted from our simulation, as shown in Fig. 2, illustrates the agent and
task’s initial positions and their aims as cited in the table above. We choose a network
from three rows and three columns, the results extracted from the first pdf document
are outlined in the next table, and the simulation presents how the agent could bide
for each task without failures (Table 3 and Fig. 3):
4) About scenario_2_part_ag:
The second scenario is developed to test the use of two agents at the same time, with
three tasks.
The main goal of this scenario is to demonstrate the capability of the agents to
bid and contract for the tasks, without interfering in each other job, and the contract
came at a specific time, depending on that time, one of the agents took the task and
provide bids, so it could be completed, depending on its priorities, rest time and
speed (Table 4 and Table 5, Fig. 4).
5) Task used in the SECOND scenario:
The grid extracted from the second simulation of the second scenario is wider and
more complicated than the first one. It is clear that in our first work, we choose to
optimize only the movement of the agents inside the grid network, but it is known in
every smart-grid, agents been hardware or software they need to move between the
components of that grid, so came the importance of our work which could facilitate
this job for the agents (Table 6, Fig. 5).
3 Discussion
At is shown in the tables below, extracted as a simulation for the different bids, our
agents have contracted for the three tasks in every step time of the first and second
scenario, the ECNP allowed them to provide many movements for each task to help
it to get to its target easily and without forgetting any of them.
We chose these particular scenarios to show that MAS are perfectly capable of
managing any size of smart grid starting with three rows and columns and ending
with a much bigger number of elements.
Using his priority, resting time, and capacity, which are variable and could be
changed from one scenario to another, the agent has the capability of managing very
well movement inside a smart grid network, without interfering in the other agent
job, executing orders and tasks in the disposal that he has been given and the most
important without forgetting any.
The simulations show that the bides are very clear and separated from each other,
which explain that the speed, capacity and rest time for every agent is respected
in both scenarios, and every bides come at a specific time and don’t collapse with
another, this property is very important of good work of a smart-grid.
4 Conclusion
The application of the extended version of contract net protocol for the movement
of agents in a smart grid of different sizes, and with a more challenging number
of agents and elements, has shown the ability of agents to manage objectives and
accomplish tasks, which can be used as a technique to forecast energy production
inside a wind farm or a photovoltaic field.
In other words, MAS are capable of achieving real-time management of distributed
systems as a microgrid, creating from it a smart-grid with self-reflections every
moment of the day, using the proprieties of MAS mentioned above, it can provide
Management of Multi-agents in a Smart Grid Network ... 1325
energy for a very large number of customers, also it can maintain its secure level of
energy by using various types of RES it can acces to.
References
16. Liu G, Jiang T, Ollis TB, Zhang X, Tomsovic K (2019) Distributed energy management
for community microgrids considering network operational constraints and building thermal
dynamics. Appl Energy 239:83–95. https://doi.org/10.1016/j.apenergy.2019.01.210
17. Sermakani S, Thangaraja M (n.d.) Power demand optimization in smart grid via wireless
networks. J Electr Electron Eng (IOSR-JEEE), 18–23
18. Ghiani E, Serpi A, Pilloni V, Sias G, Simone M, Marcialis G, Armano G, Pegoraro P (2018)
A multidisciplinary approach for the development of smart distribution networks. Energies
11(10):2530. https://doi.org/10.3390/en11102530
19. Worighi I, Maach A, Hafid A, Hegazy O, Van Mierlo J (2019) Integrating renewable energy
in smart grid system: Architecture, virtualization and analysis. Sustain Energy Grids Netw
18:100226. https://doi.org/10.1016/j.segan.2019.100226
20. Kamdar R, Paliwal P, Kumar Y (2018) A state of art review on various aspects of multi-agent
system. J Circ Syst Comput 27(11):1–5. https://doi.org/10.1142/S0218126618300064
21. FIPA standards, https://www.fipa.org/Smith R (1980) Communication and control in problem
solver. IEEE Trans Comput 29(12):12. https://www.csupomona.edu/~ftang/courses/CS599-
DI/notes/papers/contractnetprotocol.pdf
22. PankiRaj JS, Yassine A, Choudhury S (2019) An auction mechanism for profit maximization of
peer-to-peer energy trading in smart grids. Procedia Comput Sci 151(2018):361–368. https://
doi.org/10.1016/j.procs.2019.04.050
23. Smith RG (1988) The contract net protocol: high-level communication and control in a
distributed problem solver. Morgan Kaufmann, pp 357–366
24. FIPA. Fipa contract net interaction protocol specification. Technical, Report SC00029H, FIPA
(2000)
25. Bicocchi N, Cabri G, Leonardi L, Salierno G (2019) A survey of the use of software agents
in digital factories. In 2019 IEEE 28th International Conference on Enabling Technologies:
Infrastructure for Collaborative Enterprises (WETICE), 3–8 https://doi.org/10.1109/wetice.
2019.00010
26. Mařík V, Kadera P, Rzevski G, Zoitl A, Anderst-Kotsis G, Tjoa AM, Khalil I (Eds) (2019)
Industrial applications of holonic and multi-agent systems. Lecture Notes in Computer Science.
https://doi.org/10.1007/978-3-030-27878-6
27. Li K, Zhou T, Liu B, hai, & Li, H. (2018) A multi-agent system for sharing distributed
manufacturing resources. Expert Syst Appl 99(32):43. https://doi.org/10.1016/j.eswa.2018.
01.027
Smart Grid Operation Using Hybrid AI
Systems
Abstract Battery storage system is an essential component for the depletion of the
microgrid, and of the smart grid on a larger scale. It acts as an interface between
the highly intermittent renewable energy resource such as solar photovoltaic or wind
energy system, the load, such as residential homes and businesses, electric vehicle
fleets and the electric power grid. In principle, battery storage is intended to store
excess electricity production from the renewable resource. The energy stored is then
typically discharged in the evening when demand is higher, or sold to the grid when
the available electricity is higher than the demand. However, the implementation of
such a system becomes more complicated when taking into account the stochastic
nature of the renewables, the variations of the load demand and the electric vehicle
availability. In this project, a fuzzy logic management system, optimized through
genetic algorithm process, is proposed. It seeks to obtain a fully functional microgrid
that minimizes electricity consumption from the utility grid, while offering ancillary
services to the grid such as peak shaving. These objectives are accomplished by
maximizing the energy input from the renewables and optimizing the utilization
of the energy storage system. Finally, the microgrid’s control main objective is to
maintain a constant voltage at the DC bus under all operating conditions. Initial
simulations and experimental data show some promising results.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1327
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_121
1328 C. Alaoui and H. Saikouk
major challenge, especially with the inclusion of electric vehicles into the microgrid
[2–4]. The adjustment of a DC microgrid voltage with the use of classical controllers
in such environment is a challenging work and requires new and appropriate control
strategies [5, 6]. Moreover, the battery-based electric energy storage (EES) play a
key role for an efficient smart grid [7, 8] or microgrid management, in addition to
the integration of electric vehicles (EV) into the grid through an effective battery
management [9, 10].
Recently, fuzzy logic (FL) based controllers have been developed and used in
various industrial applications [11]. The performance of these controllers is influ-
enced by fuzzy rules and control coefficients that determine their level of control-
lability [12]. Compared to other adaptive control methods such as artificial neural
networks (ANN) [13], FL controllers have a special place in the industry and have
been used in a variety of applications such as electric machine ripple minimiza-
tion [12], energy management system in microgrid [14], automatic voltage regulator
system [15], load frequency control [16] and others [17, 18].
Several methods have been used to provide a solution to efficiently manage the
energy flow of a smart grid while reducing its overall costs. These methods vary from
rule-based controllers through optimization techniques to machine learning methods
with different objectives. The author of [19] proposed a rule-based controller to
dispatch a battery system with a PV and a grid-connected residential house. The
objective of the controller was to maximize the energy delivered from the PV. The
results showed a significant reduction from the power taken from the grid. The author
of [20] implemented Mixed Integer Linear Programming (MILP) for a similar system.
They used a weighted multi-objective optimization to minimize energy consumption,
costs and peak power demands. They achieved a reduced peak demand of 50%
and a reduced cost of 20%. While the authors of [21] used a similar system with
the objective of reducing the overall costs and CO2 emission. Genetic algorithms
have been used by [22, 23]. They achieved higher system efficiency and higher self-
consumption from PV panels. Many universities and research centers, such as in [24,
25], have used machine learning techniques for microgrid control. Their objective
was to maximize the usage of the battery bank and the energy input from a PV system.
The authors of [26] used 2-step ahead Q-learning system to manage a grid-tied home
having a battery pack and a wind turbine. The results of this study revealed that
while the use of the battery and wind turbine to supply the load improved by several
percent’s after the agent had learned, the economic benefit was less than 1%. Finally,
the author of [27] used a Q-learning parametrized by neural networks in order to
control a system consisting of PV, battery and hydrogen storage. They have been
able to reduce the costs of imports from the grid.
The aim of this project is to implement a fuzzy logic system, whose parameters
are optimized with genetic algorithm, that can control the charge scheduling of a
grid-tied battery, photovoltaic and wind energy residential system, with the objective
of taking the minimum energy from the grid, while making the grid to benefit from
ancillary services such as peak demand shaving. Two scenarios will be considered:
an electric vehicle connected to the smart grid, and disconnected from the smart grid.
Finally, the validation of the proposed algorithm was done in Morocco, where the
Smart Grid Operation Using Hybrid AI Systems 1329
electricity tariff is not in hourly rate, but rather in monthly rate, where the electricity
rates vary monthly from ‘low-power’, ‘high power’ and ‘peak power’ consumption
periods [28].
2 System Description
The general block diagram of the microgrid is shown in Fig. 1. The main components
of the proposed system are a PV system (PVS) with appropriate MPPT system, and a
feedback controlled boost converter used to connect the PV system to the 750 V DC
bus. In addition, a wind power system (WPS) equipped with a Permanent Magnet
Synchronous Generator (PMSG) and appropriate converters and MPPT system. A
battery based energy storage system (ESS) connected with a bidirectional DC-DC
converter to the main DC bus. An electric vehicle (EV) connected to the DC bus
through a battery charger/discharger. Finally, an AC load and a grid connection.
Fuzzy logic systems (FLS) possess many interesting features allowing it to manage
the energy flow within a microgrid having stochastic units such as PVS and WES,
1330 C. Alaoui and H. Saikouk
in addition to a load that contains an EV. FLS is model free, rule-based approach
incorporating linguistic description of a highly non-linear system.
However, fine-tuning a FLS necessitates a trial-and-error process that may be
long, expensive and tedious. In order to establish the rules, some prior knowl-
edge or experience is needed. Sometimes, such experience or prior knowledge are
unavailable.
Genetic algorithms (GAs) are sometimes used to optimize the FLS rules.
Depending on the available information and on the system’s settings, GA can be
used to optimize the fuzzy rules and/or the fuzzy memberships. Fig. 2. shows the
block diagram of such a system.
Conventional photovoltaic modules are typically rated between 200 and 300 W with
an open circuit voltage range of 20–30 V. For the PV-grid charging system, the
modules are arranged in series-parallel strings to attain the required working voltage,
current and power. DC-DC converters are connected as an interface between the PV
array and the DC bus. In order to transfer the maximum power to the DC bus at all
times, two converters are used as shown in Fig. 3. The first converter, referred to as
the boost DC-DC converter, tracks the MPPT of the PV panels in real time. At its
output, the maximum power is transferred, but with varying output voltage value,
which correspond to different irradiation and temperature values. This is shown in
Fig. 4. In order to maintain a constant voltage value at the output of the PV system,
a second DC-DC converter is added, and with appropriate feedback control, in order
to maintain 750 V at the output and hence connect it to the DC Bus. The switching
ON/OFF of the PV panels to the grid is decided by the second stage converter; if
Smart Grid Operation Using Hybrid AI Systems 1331
Fig. 3 PV panel connected to the DC bus through a first stage converter for the MPPT, and the
second stage to maintain 750 VDC at the output
the voltage input to the second inverter is below a threshold V PV,th such that the 750
V cannot be established at the output DC-DC converter and the feedback controller,
the fuzzy control system will disconnect the PV system from the DC bus.
The wind power system (WPS) consists of a Permanent Magnet Synchronous Gener-
ator (PMSG) rotated by a wind turbine which is interfaced to the DC microgrid
through a proper AC-DC power converter. Eq. (1) shows the mechanical power PM
available at the shaft of the turbine.
1332 C. Alaoui and H. Saikouk
where ρ is the air density (kg/m3 ), A is the cross-sectional area of the stream (m2 )
and V w is the free wind velocity (m/s). For a small wind turbine with fixed pith angle,
C P is solely a function of the tip speed ratio λ expressed in (2).
ωr ot R
λ= (2)
VW
where, R is the radius of the rotor (m) and ωrot is the rotational speed of the rotor
(rad/s). A typical C p -λ curve is shown in Fig. 5.
Incremental MPPT technique was selected for this wind turbine. Its main advan-
tage is its independence from the wind turbine parameters. This algorithm solely
requires power output measurement for operation. This can be readily achieved from
the DC link current and voltage without requiring any speed or wind measurements.
Fig. 6. shows the algorithm of the MPPT. It uses the perturb and observe principle
to increase the delivered power by increased the duty cycle of the PWM signal sent
to the boost converter of Fig. 7.
The battery ESS can only be charged whenever there is a surplus power from
the PVS or WTS. The state of charge (SOC) of the ESS is expressed in Eq. (3)
as [26, 29].
Smart Grid Operation Using Hybrid AI Systems 1333
n
n
S OC(t) = S OC(0) + ηc Pch,i(t) − ηd Pdis,i(t) (3)
t=1 t=1
where ηc and ηd are the charging and discharging efficiencies of the battery pack,
Pch,i(t) and Pdis,i(t) are the power for charging and discharging the batteries at time t.
Fig. 8 shows the block diagram of the battery based energy storage system (ESS).
The DC-DC converter maintains the voltage output at 750V as long as the SOC of
the batteries allows it.
1334 C. Alaoui and H. Saikouk
Figure 9 shows the bidirectional DC-DC converter that will act as a battery charger,
or as a battery discharger that will transfer the battery pack energy into the grid
through the DC bus.
3 Microgrid Operation
In order to establish the appropriate rules for the operation and control of the proposed
microgrid, the following guidelines were established:
i. The ESS and the EV can be recharged according to two modes: ‘NORMAL’
or ‘URGENT’
ii. Either the ESS or the EV can be charged in ‘URGENT’ mode, they cannot be
operating in this mode simultaneously
iii. The fuzzy controller reads the voltage levels at each component in order to
estimate its power and energy status
iv. The EV cannot be recharged from the ESS. This would incur important energy
losses due to the cycling efficiencies of the batteries.
v. The power demand from the load PL must be met at all times.
For the proposed microgrid, there are seven possible modes of operation, as
described below and illustrated in Fig. 10. In each mode of operation, the fist equa-
tion expresses the condition necessary to pass to the mode of operation as in Eqs. (4,
6, 8, 10, 12, 14, 16), and the second equation shows the power and hence the energy
transferred, this is shown in Eqs. (5, 7, 9, 11, 13, 15, 17). The DC link voltage is used
as the criteria for switching between the seven modes of operation.
Mode 1: The power generated by the renewable sources and the power available in
storage devices BESS and EV is higher than the power demanded by the load (SOC EV
> 50%). In addition, this mode of operation matches peak power demand from the
grid. Satisfying the Grid peak demand and load demand, EV being connected and
Smart Grid Operation Using Hybrid AI Systems 1335
discharging. The condition to pass to this mode is expressed in (4), and the power
received by the grid is (5).
Mode 2: This mode of operation is similar to mode 1, except that the EV needs to
be recharged (SOC EV < 50%). Satisfying the Grid peak demand and load demand, EV
being connected and charging. The grid receives power expressed in (7) as ancillary
support.
Mode 3: This mode of operation is also similar to mode 1, except that the EV is
disconnected from the microgrid. Satisfying the Grid peak demand and load demand,
EV being disconnected
Mode 4: When enough power is received from the renewable sources and the
BESS is below a threshold (SOC ESS < 50%), the available energy is used to recharge
the BESS assuming that there is no peak demand from the grid. Moreover, the EV
is not connected in this mode of operation. Recharging ESS and load demand, EV
being disconnected
Mode 5: When the BESS is almost fully discharged (SOC ESS < 10%), it must
include the power from the grid in order to reach SOC ESS = 50%. Beyond 50% of
recharge, the BESS is recharged in mode4. Recharging ESS in urgent mode, EV
being disconnected
Mode 7: When the energy available in the EV is below a threshold (SOC EV <
10%) and very little to no energy input from the renewables, then the EV should be
recharged in urgent mode by using power from the grid. Recharging EV in urgent
mode
4 Microgrid Simulation
The system under study consists of a grid-connected house situated in the north
of Morocco. It is equipped with photovoltaic panels (PV) rated at 7 KW, a wind
Smart Grid Operation Using Hybrid AI Systems 1337
5 Conclusion
This paper has presented the initial results of a FLC-based design of an energy
management system for smoothing the voltage level of a DC microgrid. It contains
renewable resources such as PV and Wind power systems, an EV, an energy storage
system and a grid-connected load.
The proposed approach is implemented on a microgrid simulation model to vali-
date the proposed methodology. Simulation of real-condition operation including
experimental data has been performed. The preliminary results of this study has
shown that it is possible to train an RL agent using a continuous action algorithm
to control the charging scheduling of a battery in a residential setting. Initial results
showed the feasibility of this agent, although the reward function needed further
tuning.
Smart Grid Operation Using Hybrid AI Systems 1339
References
21. Pena-Bello A et al (2017) Optimizing PV and grid charging in combined applications to improve
the profitability of residential batteries. J Energy Storage 13:58–72
22. Matallanas E et al (2012) Neural network controller for active demand side management with
PV energy in the residential sector. Appl Energy 91(1):90–97
23. Glavic M, Fonteneau R, Ernst D (2017) Reinforcement learning for electric power system
decision and control: past considerations and perspectives. IFAC-Papers On Line 50(1):6918–
6927
24. Leo R, Milton RS, Sibi S (2014) Reinforcement learning for optimal energy management
of a solar microgrid. In: Global Humanitarian Technology Conference-South Asia Satellite
(GHTC-SAS), pp 183–188. IEEE
25. Kuznetsova E et al (2013) Reinforcement learning for microgrid energy management. Energy
59:133–146
26. François-Lavet V, et al (2016) Deep reinforcement learning solutions for energy microgrids
management. In: European Workshop on Reinforcement Learning (EWRL 2016) (2016)
27. Schaarschmidt M, Kuhnle A, Fricke K (2017) Tensor force: a tensor flow library for applied
reinforcement learning. Web page https://github.com/reinforceio/tensorforce
28. Website: electricity tariff in morocco https://www.invest.gov.ma/index.php?Id=34503&lan
g=en. Accessed 19 Jul 2020
29. Sassi HB, Errahimi F, Essbai N, Alaoui C (2018) A comparative study of ANN and
Kalman Filtering-based observer for SOC estimation. In: IOP Conference Series: Earth and
Environmental Science vol 161, p 012022
Economic Comparison Between Two
Hybrid Systems (Wind-Hydrogen)
and (Wind-Hydroelectric) for Electricity
Production in Socotra, Yemen
Abstract Renewable energy sources are one of the main sources of energy produc-
tion, Therefore, all researchers interesting in these sources, and consider it as a
primary source to cover a country’s needs, especially in developing countries such
as Yemen, which suffer from a large deficit to cover region’s need for electric
energy, including remote islands such as Socotra. In this paper, we will present
an economic study for electricity production by wind turbines in Socotra Island, and
an economic comparison between two means of energy storage, which is energy
storage by hydrogen production or by hydroelectric, and we accurately estimate the
energy unit price (kWh) to know the economic cost of two storage methods.
1 Introduction
Renewable energy sources have grown significantly in recent times due to the prob-
lems that fossil fuels cause to environment such as air pollution, climate change, and
other problems, which have made all researchers and scientists a trend to seriously
study renewable energy and classify it as an energy that ensures a balance between
energy security and economic development [1]. Globally, The annual growth rate
of wind energy is about 30% annually, this is an indication that many countries
are interested in this source type to energy production and reducing dependence on
fossil energy [2]. In remote areas that are far from the power grid of main station, the
demand for energy sources increases, especially wind energy, and since wind energy
changes seasonally, this energy must be stored, one of the best storage methods is
the method that clean, renewable, and hybrid to energy produce at need time, peak
energy use, and use Part of the excess wind energy, especially in off-peak, to store
it as hydroelectric energy, which has been greatly distinguished recently due to its
high capacity and ease of use [3], or to produce hydrogen to be used when needed to
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1341
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_122
1342 S. Serag et al.
electricity production by fuel cells, which has proven high efficiency to produce elec-
tricity better than other types of fuel [4], and it will noted in this study the economic
difference between these two methods according to the knowledge of energy unit
price for each of storage methods separately [2].
Yemeni population is considered to have the least access to electrical energy, as
the population rate benefiting from electricity is 51.7% of total population number
[5], due to Yemen circumstances which have passed from 2011 until today, total
capacity of energy production has decreased to 70%, which has negatively affected
of Yemeni population significantly [6]. All this makes it necessary to plans study
to energy production by renewable sources, specifically wind energy, as the average
wind speed for coastal areas in Yemen is 8 m/s [7], which is sufficient to energy
production with high efficiency. This study has been applied in Socotra island as a
field of study, which is one of the largest Yemeni islands, and contains a population
density of approximately 80,000 people, which increase their need for electricity
energy recently [8], and it will check the economic cost of producing and storing
the energy by renewable energy sources, and compare it in cost, which has great
importance, as it gives a future vision of possibility of renewable energy using,
storing it by a clean, environmentally friendly methods, and even at lowest costs, as
the economic comparison enables to know the best methods of electricity production,
least cost, and best in efficiency.
2 Site Data
Socotra is Yemen island, and considered within the Hadramawt governorate, located
250 km east of the Africa Horn and 350 km south of Arabian Peninsula [8]. It has an
area of 3,600 m3 and height above sea level is 257 m. All remote islands suffer from
Economic Comparison Between Two Hybrid Systems … 1343
3 Methodology
Wind energy cost is calculated by knowing several factors, turbine type used, its
capacity factor, tower height, annual amount of energy produced according to wind
speed, its distribution during year, and lifetime for energy production, then applying
the following equations [1, 10]:
1+i 1+i t 1+i t
P V C = Cinv + Comr × × 1− − Cs × (1)
r −i 1+r 1+r
Where Cinv investment cost, Comr Operation and maintenance costs, Cs Additional
costs, i interest rate = 1.5%, t Turbine lifetime = 20 years, r inflation rate = 2%.
PV C
(kWh cost) = . (2)
PE
Where PE = energy production in a rated lifetime.
This study will be applied by turbines (Nordex 1500 kW) with height towers
between 65–80 m and cost approximately 1220 $/kW [11].
LCOE. (Levelized cost of electricity generation) which is equal to total cost divided
by total energy produced during the lifetime, for renewable energies calculated by
formula [12–14]:
1344 S. Serag et al.
n It +Mt +Ft
t=1 (1+r )t
LC O E = n Et
(3)
t=1 (1+r )t
Storage. There are several methods for hydrogen store according to hydrogen state
(solid, liquid, gas) and each method has its advantages and disadvantages, but the
best economic way is to store hydrogen in a gaseous state under a pressure of 700 bar,
it is less expensive, simpler than other methods. We will note in Table 2 Storage cost
per kg of hydrogen, percentage of costs allocated to compression, operation, and
maintenance [17].
Electricity Production (Fuel Cell). When the amount of energy generated from
turbines decreases due to peak use or low winds in region, hydrogen stock must
be used to electricity production to cover the electricity need. Fuel cells are used to
convert hydrogen into electricity in oxygen presence and the reaction result is electric
current and water. Table 3 shows the price of 1 kilowatt of fuel cell, its efficiency,
percentage of operation and maintenance during its lifetime [19].
Transport Cost. In event that hydrogen needs to be transported from the generation
place to the consumption place, there are many ways to transport hydrogen, the best
way is tube trucks, and the transportation costs will appear in Table 4, [20].
Fig. 2 Hydrogen production and storage stages for Using to electricity production
stations with storage range from 1050 $ to 7650 $ per kW, depending on the efficiency
and use period of the station. Table 5 shows the potential price of hydroelectric
stations with storage, along with cost of maintenance, operation, and lifetime [14].
Figure 4 shows the monthly average wind speeds for nine years from 2006 to 2014,
and clearly shows a large amount of wind speed, especially in June, July, August,
and September, while average of wind speed value was in January and December,
slightly lower values in remaining months, the lowest value for that speed is in April
4.37 m/s and the largest value in July is 17.22 m/s. As for the mean of wind speed for
9 years as a whole, it reached 9.17 m/s, a value that emphasizes a large and sufficient
amount to use for energy production.
When using a simulation program such as ALWIN, and by adding wind speed
data for area and type of turbine used (Nordex 80−1500 kW), the results are shown
in Fig. 5.
Economic Comparison Between Two Hybrid Systems … 1347
17.22 15.91
20.00
Wind Speed
13.99
11.04
8.37 7.91
10.00 6.79 5.63 6.64 5.65 6.54
4.37
0.00
Nov.
dec.
Apr.
Aug.
Jan.
May.
Jun.
Jul.
Oct.
Feb.
Mar.
Sep.
Month
It is noted that all variables related to wind speed, such as Weibull parameters (a
= 13.55 m/s, c = 1.42), average speed at a height of 80 m (vm = 12.34 m/s), as well
as the variables for turbine, which are average power (Pm = 635.6, 859.1 and 677.8)
kW, annual energy generated by the turbine (5567.7, 7525.8, and 5937.7) MWh, and
capacity factor (42, 56.9, and 45) %, were calculated in three cases: Measured values,
Rayleigh distribution, Weibull distribution.
By taking average annual energy for three values, measured, Weibull and Rayleigh
will be 6343 MWh, knowing the island’s population, and amount of energy consumed
by Yemeni citizen, we can know the number of turbines needed to cover the region’s
need for electricity, as shown in Table 6.
Thus, the number of turbines needed to cover island’s energy need is 3 turbines
with an annual excess of energy of 1429 MWh, equivalent to 3915 kWh per day, and
this excess energy contributes in storage process for use when needed.
1348 S. Serag et al.
By applying Eqs. (1), (2), calculating values of (Cinv , Comr., Cs ), and compensating
for parameters ( i = 1.5%, t = 20 years, r = 2%), it get the results as in Table 7,
where it was found that the total cost of energy producing by wind turbines is (3.24
× 107 )$ and energy produced from those turbines during their lifetime is (3.8 × 108 )
kWh, Thus, we can see the cost of kWh equal to (0.085) $.
By knowing cost and values , which appear in Tables 1, 2, 3, and 4, and applying the
equation of total cost to total energy production in a lifetime, we can get the results
shown as follows:
Table 8 shows the cost for each stage of hydrogen production in dollars per kWh,
and Table 9 shows total costs with or without transportation (without transportation
in the case of direct electricity production so that the hydrogen production plant and
energy generator plant are adjacent and do not need to transport the hydrogen from
one place to another), as it is shown that the largest costs are for the production
Economic Comparison Between Two Hybrid Systems … 1349
and storage of hydrogen because of the large energy used for production or storage
under pressure of not less than 700 bar. While the cost of transportation occupies a
relatively large percentage of 21% for same reason, (energy required for pressure,
storage, and cost for truck tanks and pipes, which are used for transportation).
After studying the costs of electric energy generating from wind turbines, storage
methods, and generation of electricity by (hydrogen production and storage) or
by (hydroelectric storage), we ought to study the total costs of hybrid systems
that link (wind-hydrogen) or (wind-hydroelectric) energy, or even the use of three
systems Wind, hydrogen and hydroelectric. Table 11 shows the total costs for
(wind-hydrogen), (wind-hydroelectric), and (wind-hydrogen-hydroelectric).
It is noticeable the total costs shown in previous tables are low compared to
electricity prices resulting from oil derivatives, especially with high prices and even
lack of in recent times in Yemen. Therefore, electricity production in these ways is
lower in cost, safer, more sustainable, and it is recommended to be used.
1350 S. Serag et al.
5 Conclusion
Renewable energy sources are the best and cheapest way to electricity production,
including wind energy, but because of the instability of these sources, energy must be
stored for use when needed, so this study was applied on the Socotra island, Yemen
for electric energy production by wind energy and store this energy in two methods,
first is produce, store hydrogen, and second is hydroelectric, these methods have
been economically compared and we get the following:
1. Socotra Island has a mean wind speed of 9.17 m/s. This result of the average
winds for nine years, which is a great speed that makes this island able to produce
electricity by turbines with high efficiency.
2. When using turbines Nordex 1500 kW with tower height equal 80 m, we find
an annual average energy produced from this turbine is 6,343 MWh, since the
annual consumption of the island is 17,600 MWh, Island needs three turbines
to cover regions by electricity with energy excess approximately 1429 MWh.
3. The unit cost of energy produced from wind turbines, according to the parame-
ters concluded from wind distribution and its probability is approximately 0.085
$/kWh.
4. Electricity producing process from hydrogen goes through several stages, the
most important of which is hydrogen production and storage, cost has been
calculated for all of those stages and it has been found that the total cost is equal
to 0.054 $/kWh to electricity production from hydrogen without transportation
and 0.068 $/kWh by transportation.
5. Hydroelectricity is the easiest and safest method for electricity production, the
costs of electricity producing by it are 0.043 $/kWh, which is less expensive
compared to other storage and production methods.
6. When applying a hybrid system (wind-hydrogen) and (wind-hydroelectric) or
(wind—hydrogen and hydroelectric), it is found that the costs are ((0.14 or 0.15),
0.13, 0.18) $/kWh respectively, which are low costs compared to the methods
of producing electricity with fossil fuels, especially in recent times when fuel
value has increased.
Economic Comparison Between Two Hybrid Systems … 1351
References
1. Ohunakin OS, Akinnawonu OO (2012) Assessment of wind energy potential and the economics
of wind power generation in Jos, Plateau State, Nigeria. Energy Sustain Dev 16(1):78–83
2. Shawon MJ, El Chaar L, Lamont LA (2013) Overview of wind energy and its cost in the Middle
East. Sustain Energy Technol Assess 2:1–11
3. Kumar K, Ansari MA (2013) Design and development of hybrid wind-hydro power generation
system. In: 2013 International Conference on Energy Efficient Technologies for Sustainability,
Nagercoil
4. Garland NL, Papageorgopoulos DC, Stanford JM (2012) Hydrogen and fuel cell technology:
progress, challenges, and future directions. Energy Procedia
5. Alkholidi AG (2013) renewable energy solution for electrical power sector in Yemen. Int J
Renew Energy Res (IJRER), 9
6. Saleh Qasem AQ (2018) Applications of renewable energy in Yemen. J Fundam Renew Energy
Appl 08(02) (2018)
7. Hashim Alkipsy EI, Raju V, Kumar H (2020) A review of the challenges of energy sector and
prospects of renewable energy utilization in Yemen. Glob J Manag Bus
8. Fritz M, Okal EA (2008) Socotra Island, Yemen: field survey of the 2004 Indian Ocean tsunami.
Nat Hazards 46(1):107–117
9. Saqqaf A, Alkaf, Muqbel MA, Bin SM (2010) World renewable energy congress XI cost
analysis of renewable energy for power supply for remote area in Yemen: a case study for
Socotra Island
10. Soulouknga MH, Oyedepo SO, Doka SY, Kofane TC (2020) Evaluation of the cost of producing
wind-generated electricity in Chad. Int J Energy Environ Eng 11(2):275–287
11. ‘Renewable Energy Technologies Cost Analysis Series Wind Power’ International Renewable
Energy Agency IRENA June 2012
12. Kornbluth K, Greenwood J, Jordan E, McCaffrey Z, Erickson PA (2012) Economic feasibility
of hydrogen enrichment for reducing NOx emissions from landfill gas power generation alter-
natives: a comparison of the levelized cost of electricity with present strategies. Energy Policy
41:333–339. https://doi.org/10.1016/j.enpol.2011.10.054
13. Shea RP, Ramgolam YK (2019) Applied levelized cost of electricity for energy technologies
in a small island developing state: a case study in Mauritius. Renew Energy 132:1415–1424.
https://doi.org/10.1016/j.renene.2018.09.021
14. Gielen D (2012) Irena renewable energy technologies cost analysis series hydropower
15. A Grant Thornton and Clean Energy Pipeline initiative (2018) Renewable energy discount rate
survey a grant thornton and clean energy pipeline initiative, p 36
16. Chan A, Hart D, Lehner F, Madden B, Standen E (2014) Fuel cells and hydrogen Joint
undertaking, “Development of Water Electrolysis in the European Union”, February 2014
17. Riis T, Hagen EF, Preben, Vie DJS, Ulleberg Ø (2005) International energy agency hydrogen
production and storage © OECDIE
18. Touili, S Alami Merrouni A, Azouzoute A, El Hassouani Y, Amrani A (2018) A technical and
economical assessment of hydrogen production potential from solar energy in Morocco. Int J
Hydrog Energy 43(51):22777–22796
19. Fuel Cell Technologies (2016) Office of energy efficiency and renewable energy - section 3.4
20. Amos WA (1999) Costs of Storing and Transporting Hydrogen NREL/TP-570–25106, ON:
DE00006574, 6574
A New Hybrid System Used to Connect
PV to a Microgrid
Abstract This paper proposes a hybrid system by connection the photovoltaic panels
to a microgrid through an active power filter (APF). This new hybrid system enables
the limitation of the power generated by photovoltaic panels and follow the load
demand without supplying power to the microgrid. This whole unified hybrid system
will allow the attenuation of the harmonic injected to the microgrid and the compen-
sation of the reactive power in the case of lack of power from the PV system. The
Matlab/Simulink simulation environment is used to demonstrate the proper func-
tioning of this hybrid system. A modified perturb and observe algorithm is used to
control the boost converter and indirect method for the APF control.
1 Introduction
The increase of clean energy demand involves the need of using renewable sources
like PV panels. The storage of energy through batteries for later uses leads to an
increased investment costs, maintenance costs and chemical pollution associated
with batteries, which makes this option less feasible. Injecting energy directly into
the microgrid ensures the maximum use of clean power generated by renewable
energy sources [1]. The integration of renewable energy systems is the most current
trend in the field of energy research. The increase in renewable energy production
and injection into electric its grids has begun to affect the security and stability of
the operation of power systems [2]. Obtaining the maximum power from renewable
sources without affecting the grid management and the attenuation of harmonics in
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1353
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_123
1354 A. Dusa et al.
the distribution grid is the main objectives of this paper. The maximum power point
(MPP) principle and supply of nonlinear load from PV is presented and demonstrated
in the papers [3–6].
In this paper it is proposed to interconnect the PVs to the low power grid through
the shunt active power filter (APF) and injecting only the amount of power needed
by the electrical loads. The APF consists of a three-phase inverter with three levels
in Voltage Source Inverter (VSI) with two capacitors on the DC side, with the aim
of attenuating the harmonics introduced in the grid from nonlinear tasks as well as
compensating the power factor. The control algorithm of the APF is performed using
the indirect control method. The connection of the PVs with APF is made through a
boost converter. The connection between APF and boost converter is made on the DC
side. Extraction of the maximum power from the PVs and its limitation is performed
using a modified Perturb and Observe (P&O) control algorithm.
Figure 1 presents in detail the proposed diagram for this new hybrid system that
allows supply the nonlinear load, PVs output power limitation at load level and
harmonics attenuation as well as correction of the power factor.
The boost converter allows the output voltage to be raised. The wiring diagram of a
boost converter is present in Fig. 2 For designing this type of converter it is necessary
to know the following values: minimum input voltage U in_min = 345.4 V, output
voltage U out = 750 V, intensity of the load current I out = 9.25 A, switching frequency
F sw = 25000 Hz, power P = 7300 W, effectiveness η = 95%, the percentage from
the output voltage for calculating the ripple voltage is equal to 1 [7].
The design elements of the boost converter are:
The filling factor D for the PWM control signal for IGBT in the boost converter
circuit is determined by:
Uin_min · η
D = 1− % (1)
Uout
P
Iout = (2)
Uout
The value of the input current I in results from the formula: P = I in · U in = I out · U out
P
Iin = (3)
U1n
The value of the input current 1 is equal to the current rate in the coil. The ripple
current of the inductance L will be equal to 10% of the input current.
Uin · D
L= (5)
I L · Fsw
The value of the ripple voltage U out will be equal to 1% of the value of the output
voltage:
Iout · D
C= (7)
Fsw · Uout
Uout
R= (8)
Iout
The modified P&O control method [8] allows the extraction of the maximum power
from the PVs and then compares it with the load need. If the power value of the PVs
is higher than the load need then the P&O control algorithm increases the duty cycle,
otherwise is decreases the duty cycle. The diagram of the modified P&O control
algorithm is present in Fig. 3.
Indirect control is a strategy that does not require knowledge of electrical current
spectrum absorbed by the nonlinear load. This strategy imposes that the grid current
to be sinusoidal and in phase with the voltage [9].
1
where: i La (t) - fundamental active component of load current;
i Lak (t) - harmonics sum of load current;
k
i Laq (t) - reactive component of load current.
The electric current through the APF on phase a will be:
i sa = i La
1
+ i 1f a + ĩ (13)
The signal is generated for the charging input of the current regulator of phase a
of the power supply:
va
i a∗ (t) = ε DC √ = ε DC sin ωt (14)
2V
where: V is the effective value of the phase voltage of the grid and εDC the output of
the DC voltage regulator.
The above charge is compared with the measured value of the electric current
absorbed from the isa grid, resulting for the control of the APF on phase a:
1358 A. Dusa et al.
u ∗ca = k i a∗ − i sa = k i a∗ − ĩ − i La
1
− i 1f a (15)
i a∗ = i La
1
+ i 1f a (16)
so, the APF will generate the polluting component itself necessary for the nonlinear
load.
The assessment εDC from the regulator for charging capacitor C is converted to
current reference, as follows:
⎧ ∗
⎪ i = ε · sin ωt
⎨ a
i b∗ = ε · sin ωt − 2π 3 (19)
⎪
⎩ ∗
i c = ε · sin ωt − 4π 3
Sinusoidal signal in phase with fundamental are obtained with Phase-Locked Loop
(PLL), which allows the synchronization of the compensation signal and the voltage
of the power supply system.
Figure 4 presents the control scheme of the APF with the indirect method. Where
we can see the PLL block for determining the grid phase (ωt) and the DC regulator
∗
for maintaining the voltage at the set value (Vdc ).
3 Simulation Results
• PVs output power was less than the nonlinear load need and the deficit was
supplemented from the microgrid.
• PVs output power was limited at load consumption level.
In Table 1 are presented the values used for the simulation of the proposed hybrid
system.
1360 A. Dusa et al.
Figure 5 presents (a) power, (b) voltage and (c) current from the boost converter,
were:
• The first hypothesis is presented in the range 0.22–0.25 s:
– PVs produce the maximum power, PPV = 6700 W (due to losses on the boost
converter and APF);
– nonlinear load power PL = 13700 W;
– the power deficit was supplemented from the microgrid.
• The second hypothesis in the range 0.25–0.32 s:
– the PVs output power was limited to the nonlinear load need (PL = 3000 W).
In Figs. 6 and 7 is presented the voltage and current waveforms taking into account
the two hypotheses. Here is presented for the first hypothesis the waveform of the
microgrid voltage (Fig. 6(a)), the current generated by the PVs through the APF
(Fig. 7(a)) and injected in the common connection point (CCP) to supply the nonlinear
load (Fig. 7(b)) and attenuate the harmonics in the microgrid. And the current from
the microgrid (Fig. 6(b)) supplements the supply of the nonlinear load.
In the second hypothesis it is notice that the current from the microgrid is 0 (zero)
(Fig. 6(b)). This is due to the limitation of the power generated by the PVs, injecting
the power needed by the nonlinear load.
A New Hybrid System Used to Connect PV to a Microgrid 1361
Also, here we can observe the transient regime at the moment t = 0.25 s, at the
change of the power of the nonlinear load.
Figure 8 present the Fast Fourier Transform (FFT) analysis for the source voltage
waveforms on phase 1. Figure 8(a) presents the harmonic level of the source voltage
for the first hypothesis, THDu = 4.01%. Figure 8(b) presents the harmonic level of
the source voltage for hypothesis two, THDu = 4.83%.
Figure 9 present the FFT analysis for current waveforms for phase 1. Figure 9(a)
presents the harmonic level of the nonlinear load current, THDi = 29.24%.
Figure 9(b) presents the harmonic level of the microgrid current, THDi = 4.88%.
1362 A. Dusa et al.
Fig. 8 FFT analysis of the voltage waveform a) in the first hypothesis, b) in hypothesis two
Fig. 9 FFT analysis of the current waveform a) the nonlinear load, b) the microgrid current
4 Conclusions
The aim of this paper was to create a new hybrid system were to connect the PVs to the
microgrid using a shunt APF. Which will allow the maximum power to be extracted
from the PVs or to limit the output PVs power, without introducing the excess power
into the microgrid. At the same time, this hybrid system allows the attenuation of the
harmonic level in the microgrid and the compensation of the reactive power.
For this paper, simulations were performed in Matlab/Simulink, which demon-
strate the advantages of this proposed system.
A New Hybrid System Used to Connect PV to a Microgrid 1363
References
1. Ali Q, Al-Shetwi MA, Hannan KPJ, Mansur M, Mahlia TMI (2020) Grid-connected renewable
energy sources: review of the recent integration requirements and control methods. J Clean Prod
253(119831):1–17
2. Sriranjani R, Jayalalitha S (2018) PV interconnected shunt active filter for power balancing and
harmonic mitigation. Int J Pure Appl Math 118(18):2341–2354
3. Djeghader Y, Chelli Z, Rehaimia S (2018) Harmonic mitigation in electrical grid connected
renewable energy sources. ACTA Electrotehnica 59(4):287–291
4. Manimegalai V, Senthilnathan N, Sabarimuthu M (2017) Grid interfaced renewable energy
source to improve power quality using SAPF. Int Res J Eng Technol 4(4):834–839
5. Chaithanakulwat A (2019) Multi-functionality control, power filtering single-phase grid-
connected photovoltaic system. Am J Electr Power Energy Syst 8(2):62–70
6. Bouzelata Y, Kurt E, Chenni R, Altın N (2015) Design and simulation of a unified power quality
conditioner fed by solar energy. Int J Hydrogen Energy 40(44):15267–15277
7. Livinti P, Ghandour M (2020) Fuzzy logic system for controlling of DC/DC boost converter
developed in MATLAB SIMULINK. Int Res J Eng Technol 7(9):1629–1636
8. Zala MP, Pandya MH, Odedra KN, Patel DP (2017) Active power control of PV system in MPPT
and CPG mode. Kalpa Publ Eng 1:270–277
9. Gurguiatu G Munteanu T (2012) Indirect control in active power filters. In: International
Conference on Optimization of Electrical and Electronic Equipment (OPTIM) 2012 - 13th,
ResearchGate
Coordinated Control and Optimization
Dispatch of a Hybrid Microgrid in Grid
Connected Mode
Abstract This paper presents a novel daily energy management system for coordi-
nated control and optimization dispatch of a grid connected hybrid microgrid (HMG).
The main objective is to minimize both operating cost and emissions level, taking
into account ensuring continued power balance, maximize the generation of renew-
able energy sources, optimal charging/discharging of battery storage system, and the
system constraints. The economic and emission dispatch problem has been solved
using grey wolf optimizer (GWO) algorithm, under four case studies. Then a compar-
ison with other heuristic algorithms is considered. The obtained results confirm the
potential of GWO in terms of stability, robustness, and convergence speed. Therefore,
it can be used for real time control and optimization dispatch of HMG.
1 Introduction
In the last years, with the growing electricity demand and the development of
distributed generation (DG) technology. The integration of microgrid (MG) into
the existing electricity power system is considered as a promising alternative that
provides more reliable operation, considerable cost saving and reduction in pollutant
gas emissions due to excessive energy dependency on fossil fuel [1]. However, the
intermittent generation of renewable energy sources (RESs) and difficulty of predic-
tion of load demand caused a significant mismatch between the generated power and
the required demand [2]. Therefore, an energy management system (EMS) plays a
vital role in optimizing dispatch problem and coordinated control in MG system.
Several studies have been conducted in area of MG energy management and
control.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1365
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_124
1366 M. Lagouir et al.
In [3], the authors proposed a novel optimization model to minimize both opera-
tional and pollution cost in a MG, then an advanced dynamic programming (ADP)
method is employed to solve the optimization dispatch problem. To evaluate the
system performance two case studies were considered in the presence and without
a battery storage system. In [4], the authors discuss the application of ant lion opti-
mizer (ALO) algorithm for handling the optimal economic load dispatch problem.
Considering challenging constraints such as spinning reserve constraint, ramp rate
limits, prohibited operating zones of generating units. Three case studies have been
investigated, then the results are compared with other heuristic methods to prove the
stability and robustness of ALO method. Many heuristic methods are implemented
for optimal energy management in MG, including particle swarm optimization (PSO)
algorithm [5], antlion optimizer algorithm [4, 6], artificial immune system (AIS) [6],
grey wolf optimizer (GWO) algorithm [7], and others.
The main contribution of this work is the development of a novel energy manage-
ment system, with the aim to minimize both operating cost function and emissions
level of pollutant. While taking into account system power balance, high level pene-
tration of RESs, optimal charging/discharging of battery storage system (BSS), moni-
toring state of charge (SoC) of the battery, and the system’s components power supply
constraint. The grey wolf optimizer heuristic algorithm is used in order to solve the
optimization dispatch problem. And a comparison from applying GWO algorithm
and other known heuristic methods, including PSO, novel bat algorithm (NBA) [8],
whale optimization algorithm (WOA) [9], has been performed, to show the validity
and effectiveness of the proposed strategy.
The rest of the paper is organized as follows: The system description is given in
Sect. 2. Section 3 outlines the proposed energy management approach. The results
and discussion are presented in Sect. 4. Finally, Sect. 5 concludes the work.
The proposed hybrid microgrid (HMG) system is running under grid connected
mode. Figure 1 shows the architecture of the studied HMG. The actual load demand
is primarily supplied by renewable energy sources represented by photovoltaic
panels (PV) and wind turbine (WT) unit. Three conventional energy sources (CESs)
composed of diesel generator, fuel cell unit and microturbine unit (DG, FC and MT)
are settled and used as backup to supply the load whenever the combination output
power from RESs and the energy storage system (ESS) represented by battery is not
sufficient to supply the required demand. Moreover, the HMG can exchange power
with the main grid through purchasing or selling power. Finally, all the system’s
components use various conversion devices for connection to the common bus AC.
Thus they are connected to a centralized control system dedicated to dealing with
the coordinated control and optimization dispatch problem in HMG.
Coordinated Control and Optimization Dispatch … 1367
3 Problem Statement
In this work, two contradictory objective functions are considered for minimizing:
The first objective function is the operating cost and the second objective function
is the pollutant gas emissions of Nitrogen oxide, Sulfur dioxide and Carbon dioxide
(NOx , SO2 and CO2 ), respectively. The problem can be mathematically formulated
by [10]:
min O FH M G = ω · C1 + (1 − ω) · C2 (1)
The first objective function is the operating cost of generating units, it combines four
terms representing the fuel cost, operation and maintenance cost, start-up cost and
the exchanged power cost with the main grid, expressed as [11, 12]:
1368 M. Lagouir et al.
N
C1 = {C Fi (Pi ) + O Mi (Pi ) + ST Ci + (C P E i − RS E i )} (2)
i=1
N
O Mi = K O Mi Pi (4)
i=1
The start-up cost of dispatchable generator i is expressed in term of the unit has been
off [11, 13]:
To f f,i
ST Ci = σi + δi 1 − exp − (5)
τi
According to the RESs supply limitation and the actual required demand, the HMG
may import or export power from or to the main grid. The exchanged power cost can
be represented mathematically as [11]:
C P E i = C p × max(PD − Pi , 0)
RS E i = Cs × max(Pi − PD , 0) (6)
Where C p and Cs represent the price of power purchased or sold to the main grid in
[$/Kwh].
The pollutant gas emissions cost of (NOx , SO2 and CO2 ) is expressed as a linear
function of the generated power Pi from dispatchable source i and the power provided
by main grid PGrid [12, 13]:
N
M
M
C2 = χk (E Fik Pi ) + χk (E FGridk PGrid ) (7)
i=1 k=1 k=1
Coordinated Control and Optimization Dispatch … 1369
Where χk is externality costs of emission type k. E Fik and E FGridk represent the
emission factors of generator i and main grid, respectively.
In the operation of the HMG system, the power balance should be satisfied with each
time step. It is expressed as follows:
0 ≤ PP V ≤ PP VM ppt
0 ≤ PW T ≤ PW TM ppt
PE SSmin ≤ PE SS ≤ PE SSmax
Pi Min ≤ Pi ≤ Pi Max
PGrid Min ≤ PGrid ≤ PGrid Max (9)
This work aims to design and implement a two-layer energy management system,
for daily coordinated control and optimal dispatch of HMG system. The upper
layer combines a fuzzy logic control system (FLS), used to define the optimal
charging/discharging power of BSS, the decision to turn on or off the dispatch-
able sources and the interactive mode between the HMG and the main grid. On
the other hand, the lower layer consisting of grey wolf optimizer algorithm dedi-
cated to solve the economic and emission dispatch problem in case of using the
CESs. Figure 2 shows the structure of the proposed energy management system. The
following subsection briefly describes the mathematical models of GWO algorithm.
The grey wolf optimizer algorithm was proposed by Mirjalili in 2014 [14]. The basic
idea of this algorithm was inspired by hunting technique and social hierarchy of grey
wolves. To simulate the leadership hierarchy, the grey wolves living in the same pack
are divided into alpha, beta, delta and omega. The alpha is responsible for decision-
making about hunting, sleeping place and so on. In the second level of hierarchy, we
1370 M. Lagouir et al.
found beta wolves with the mission of helping the alpha in decision-making or any
other activities related to the pack. The third level in the grey wolves’ hierarchy is
delta, in this level the wolves are responsible for watching boundaries, guarantee the
safety and taking care for the weakest of the pack. And in the last level, there are
the omega wolves that assumed to follow the wolves of the highest levels in GWO
algorithm.
Mathematically, the position of wolf is modeled as follows [7, 14]:
→ −
− → − → →
−
D = C · X P (t) − X (t) (10)
−
→ −
→ −
→ − →
X (t + 1) = X P (t) − A · D (11)
−
→ −
→
Where X and X p represent the position vector of a grey wolf and the prey,
respectively. t is the current iteration.
−
→ −
→
Finally, A and C indicate coefficient vectors, and they are calculated using the
following formulas:
−
→
A = 2−
→ r1 − −
a ·−
→ →
a (12)
−
→
C =2·−
→
r2 (13)
Coordinated Control and Optimization Dispatch … 1371
X 1 = X α − A1 · D
α , X 2 = X β − A2 · D
β , X 3 = X δ − A3 · D
δ (15)
X 1 + X 2 + X 3
X (t + 1) = (16)
3
Where X α (t), X β (t) and X δ (t) represent the position of α, β, and δ, respectively.
First, it is assumed that PV and WT units generate power without any emissions,
and with a generation limits equal to 25 kw and 20 kw, respectively. The maximum
and minimum output powers of BSS are set to 30 kw and −30 kw, respectively. And
the limit on power exchanged with the main grid is considered as 30 kw. Finally, the
maximum capacity of each dispatchable generator is set to 30 kw. Figure 3 shows
the actual output power from both RESs and optimal output power of the BSS based
on a fuzzy logic control system. As it is observed, the combination output power
of RESs and BSS is not enough to serve the load demand, therefore the fuzzy logic
controller incorporates the dispatchable sources (DG, FC, and MT) in the energy
management, while permits importing power from the main grid. Figure 4 reveals
optimal output power of CESs and main grid. And Fig. 5 shows the contribution
rates of each power generation source. As we can see, the best choice to minimize
both operating cost and emissions level is to make use of generated power of FC and
MT unit, respectively. Hence, the DG is dispatched only when the demand becomes
very high. Finally, the total system balance for a daily simulation time is shown in
Fig. 6. It is to be noted that the previously described figures are plotted with a weight
coefficient ω = 0.16, and a time step of 15 min.
To evaluate the impact of the selected weight coefficient on the formulated objec-
tive function, the operating cost and emissions level are calculated for different values
of ω, as is shown in Fig. 7. It is clearly seen that as long as ω is increased from 0
to 1, more the priority is given to minimizing the operating cost, therefore the emis-
sions level becoming higher (Fig. 8). Similarly, when the value of ω is decreased,
the emissions level are first considered to minimized. According, to the results better
compromise between the two contradictory objective functions is achieved when the
value of ω = 0.16. Finally, Table 1 summarizes the obtained results considering four
Coordinated Control and Optimization Dispatch … 1373
For all case studies, the population size is set to N = 30, and the maximum
iteration I termax is considered as 100.
6 Conclusion
In this paper, a novel approach based on combining the fuzzy logic control system
with a GWO heuristic method, is introduced to deal with the control and optimization
dispatch problem in hybrid microgrid system. Four case studies are considered, then
the system’s performance is compared with other algorithms, including PSO, NBA,
and WOA. Simulation results prove that GWO is most effective in determining
the global optimal solution in less time, which confirms the validity of the proposed
approach to efficiently manage the generated power of each component of the studied
HMG system, considering technical, economic and environmental criteria.
Coordinated Control and Optimization Dispatch … 1375
References
1. Murty Vallem VVSN, Kumar A (2020) Optimal energy dispatch in microgrids with renewable
energy sources and demand response. Int Trans Electr Energy Syst 30(5):1–27
2. Ren Sh, Wang J, Ma M (2019) Multi-objective optimal control of microgrid based on economic
model predictive control. In: Chinese Control Conference (CCC), pp 1–6. IEEE, Guangzhou
3. Moradi H, Esfahanian M, Abtahi A, Zilouchian A (2018) Optimization and energy management
of a standalone hybrid microgrid in the presence of battery storage system. Energy 147:226–238
4. Van TP, Snasel V, Nguyen TT (2020) Antlion optimization algorithm for optimal non-smooth
economic load dispatch. Int J Electr Comput Eng 10(2):1187–1199
5. Radosavljevic J, Jevtic M, Klimenta D (2015) Energy and operation management of a microgrid
using particle swarm optimization. Eng Opt 48(5):811–830
6. Hatata AY, Hafez AA (2019) Ant lion optimizer versus particle swarm and artificial immune
system for economical and eco-friendly power system operation. Int Trans Electr Energy Syst
29(4)
7. Cherkaoui N, Belfqih A, El Mariami F, Boukherouaa J, Berdai A (2020) Active power output
optimization for wind farms and thermal units by minimizing the operating cost and emissions.
Int J Electr Comput Eng 10(4):3412–3422
8. Meng XB, Gao XZ, Liu Y, Zhang H (2015) A novel bat algorithm with habitat selection and
doppler effect in echoes for optimization. Expert Syst Appl 42(17–18):6350–6364
9. Mirjalili S, Lewisa A (2016) The Whale optimization algorithm. Adv Eng Softw 95:51–67
10. Nwulu NI, Xia X (2015) Multi-objective dynamic economic emission dispatch of electric power
generation integrated with game theory based demand response programs. Energy Convers
Manag 89:947–963
11. Mohamed FA, Koivo HN (2012) Online management genetic algorithms of microgrid for
residential application. Energy Convers Manag 64:562–568
12. Nemati M, Braun M, Tenbohlen S (2018) Optimization of unit commitment and economic
dispatch in microgrids based on genetic algorithm and mixed integer linear programming.
Appl Energy 210:944–963
13. Alvarado-Barrios L, Rodríguez del Nozal A, Tapia A, Martínez-Ramos JL, Reina DG (2019)
An Evolutionary computational approach for the problem of unit commitment and economic
dispatch in microgrids under several operation modes. Energies 12
14. Mirjalili S, Mirjalili SM, Lewis A (2014) Grey Wolf Optimizer. Adv Eng Softw 69:46–61
Smart Energy Management System:
SCIM Diagnosis and Failure
Classification and Prediction Using
Energy Consumption Data
Abstract This paper represents the impact of smart energy management system
in industrial microgrid on improving the key performance indicators of the mainte-
nance, the case study will be on the squirrel cage induction motors, the idea is to
finding correlations between energy consumption data and mechanical failures of
these machines, this work shows a first version of a test bench acquiring vibration,
voltage, current and speed while the eccentricity fault is present, after test bench
preparation, installation, configuration, gathering the data, monitoring it, and see the
harmonics of current and vibration in frequential domain and see the first results, the
idea is to store the data and trying to implement machine learning algorithms, which
gives an accuracy between 0.71 and 0.96 between model features which are current
voltage and vibration, this work is a primary work of a big idea that is already devel-
oped and mention by a lot of scientists and researchers, that can be implemented in
a real case of study in mining industry.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1377
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_125
1378 O. Laayati et al.
1 Introduction
This work represents new methods for classifying SCIM faulty behaviors, by only
using current and voltage data, that means specific energy consumption data which is
acquired by a smart grid monitoring system, the goal is using new machine learning
techniques and algorithms that help finding correlations between vibration data, and
classifies failures [5], such as an unsupervised and general algorithm that can classify
an anomaly while a voltage unbalance is detected, and supervising the oscillations
of the current third harmonic.
2 Methodology
2.1 Introduction
The idea is to prepare a test bench to test the data acquisition and monitoring system,
installing and testing the current, voltage, speed and vibration sensors on the test
Smart Energy Management System … 1379
bench, monitoring the results using a data acquisition software, monitoring these
data and store it then apply classification and prediction algorithms to know if there
is some correlations between data.
The next step is to implement an industrial solution in the mining industry to test
the efficiency and measure the performance of the smart prediction and classification
algorithms and control its impact on the maintenance and production.
The main defects and failures of a SCIM showed in the state of art [6, 7] are:
The rotor eccentricity which is the most common problem in the mining industry
in conveyors and draglines, this fault is due to an offset between the axis of symmetry
and the rotation axis, we can see a lot of eccentric bearing and eccentric armature
motor.
The breakage of rotor bars and end-bearings this type of failures is classified
as critical failures in SCIM motors and can generates damage and more energy
consumption due to friction and mal functioning of the bearing lubrication.
The rotor bows this fault is due to the heating or cooling of the rotor, it causes a
local thermal asymmetric distribution. And other failures after the damage of the rotor
that can increase the energy consumption and impact voltage and current behavior.
There are many failures analysis techniques that helps to detect or to predict defects
of SCIM motors, we mention here some popular one such as acoustic emission, Air-
gap torque, stator current, electromagnetic field monitoring, induced voltage, instan-
taneous angular speed, instantaneous power, motor circuit analysis, surge testing,
vibration, and voltage.
In this work we will focus on electrical energy consumption data, which is current,
voltage, and time, and correlate these data with faulty SCIM motors vibration data
regarding torque and speed.
The voltage unbalance and electrical signature is a fault can be detected by current
signal [8] it can be applicable in framework of 3 different signal domain, time, domain
knowledge and hybrid time frequency data [9], the spectral analysis of the current of
SCIM motor using the spectrogram background theory short time fourier transform,
the auto winger distribution and the wavelet decomposition [10], for the detection
of air gap eccentricity the most popular method is the stator current monitoring, by
showing shape deviations in park’s vector which is a circular pattern normally [11],
the comparing method by combining harmonics of electrical components, power
1380 O. Laayati et al.
supply frequency and harmonics of cage frequency which is the rotation of the rotor,
getting spectral amplitude and comparing the slopes show very efficient results to
classify and detect failures [12].
Fig. 1 General scheme of test data mining, preparation, clustering, and visualization
Smart Energy Management System … 1381
3 Case Study
3.1 Introduction
This case study will discuss the feasibility of this idea and the general acquisition
system architecture, with monitored data and results, we will propose the data mining
steps data acquisition, introducing machine learning classification and prediction
models and the key performance indicators that will be monitored.
In this figure we are proposing a scheme to gather data using National Instruments
Modules of vibration, current voltage and rotation speed, the gathered data using
LABVIEW and stored into a PostgreSQL database, introducing python code to
extract data from POSTGRESQL database to check correlations between data. This
scheme is based on the intelligent power quality monitoring test bench using Internet
of Things monitoring [11].
This test bench has options to create and to simulate defects and failure, in this paper,
we will focus on the eccentricity fault detection, the acquired data, of voltage of the
stator in each line [13], the current, the root mean square value of each current line
and its Short Time Fourier Transform STFT [6], the vibration and its FFT in 4 points
axial and vertical, and the rotation per minutes.
The data is being recorded into the database, using database connection toolbox
for historical data with timestamp, extracted by python code, preprocessed, and
recalculated using predefined machine learning algorithm to find and explore visual
correlations to choose effectively good clustering, classification, and prediction
algorithms.
The Fig. 3 graph shows the RMS and the vibration temporal values with STFT,
the current, signal, RMS and the STFT in the Fig. 4, we are using these data to see if
there is any correlations and implementing a multilinear regression and see the score
of each, the Fig. 5 shows the voltage signal to detect voltage unbalance.
4 Conclusion
This work identifies correlations between STFT Current and vibration data, in the
test bench simulation and data gathering of a faulty and healthy conditions, there are
some differences in the data plots shown in the test, the voltage unbalance is not well
shown in this test bench simulation, because of the condition of the SCIM motor
which is in a good health, the machine learning basic linear regression algorithms
shows good results between STFT vibration and current of each current line, which
1384 O. Laayati et al.
Fig. 6 Correlations plots between STFT vibration and STFT current healthy condition.
Fig. 7 Correlations plots between STFT vibration and STFT Current faulty condition.
is a good indicator to continue working on this idea and implementing new machine
learning algorithms, to classify and to predict other failures and SCIM electrical and
mechanical defects and finding more correlations between all faults.
Smart Energy Management System … 1385
Fig. 8 Accuracy of linear regression model between STFT Signal of current and vibration.
This paper shows the importance of gathering energy consumption data in main-
tenance of SCIM motors, this idea can be implemented the smart micro grid in the
mining industry to control every single SCIM with current and voltage and save a
lot of time of maintenance and increase production and the profitability.
5 Future Work
In the future work we propose to continue on gathering more data, and testing all
mechanical and electrical defects of the SCIM motors in the hardware, in software
we are proposing to implement machine learning algorithms, such as neural network,
Support vector machine, dCCN, and trying to image data to better understand the
behavior of the faulty and health SCIM machine, the next step is to develop an applica-
tion and implement it in a real case study of jaw crusher in mining industry, gathering
data, and classifying its faults and problems, while controlling the Maintenance KPIs.
References
1. Singh G, Anil Kumar TCh, Naikan VNA (2018) Efficiency monitoring as a strategy for
cost effective maintenance of induction motors for minimizing carbon emission and energy
consumption. Reliab Eng Syst Safety. https://doi.org/10.1016/j.ress.2018.02.015
2. Abdelati R, Mimouni MF (2011) Analytical solution of optimized energy consumption of
induction motor operating in transient regime. Eur J Control 4:397–411 © 2011 EUCA. https://
doi.org/10.3166/EJC.17.397-411
3. Kumar P (2020) Ananda Shankar Hati: deep convolutional neural network based on adaptive
gradient optimizer for fault detection in SCIM. ISA Trans. https://doi.org/10.1016/j.isatra.2020.
10.052
4. Raja Singh R, Yash SM, Shubham SC, Indragandhi V, Vijayakumar V, Saravanan PC, Subra-
maniyaswamy V (2020) IoT embedded cloud-based intelligent power quality monitoring
system for industrial drive application. Future Gener Comput Syst 112:884–898
5. Dorrell DG, Thomson WT (1997) Analysis of airgap flux, current, and vibration signals as
a function of the combination of static and dynamic airgap eccentricity in 3-phase induction
motors. IEEE Trans Ind Appl 33:24–34
1386 O. Laayati et al.
6. Frosini L (2019) Novel diagnostic techniques for rotating electrical machines—a review. Elec-
trical machines. Design, control and diagnosis (WEMDCD), Athens, Greece, 22–23 April
2019
7. Mehrjou MR, Mariun N, Marhaban MH, Misron N (2011) Rotor fault condition monitoring
techniques for squirrel-cage induction machine—a review. Mech Syst Signal Process 25:2827–
2848
8. Ebrahimi BM, Roshtkhari MJ, Faiz J, Khatam SV (2014) Advanced eccentricity fault recog-
nition in permanent magnet synchronous motors using stator current signature analysis. IEEE
Trans Ind Electron 61:2041–2052
9. Hendrickx K, Meert W, Mollet Y, Gyselinck J, Gryllias BCK, Davis J (2020) A general
anomaly detection framework for fleet-based condition monitoring of machines. Mech Syst
Signal Process 139:106585
10. Burnett R, Watson JF, Elder S (1996) The application of modern signal processing techniques
for use in rotor fault detection and location within three-phase induction motors. Signal Process
49:57–70
11. Faiz J, Moosavi SMM (2016) Eccentricity fault detection – from induction machines to DFIG—
a review. Renew Sustain Energy Rev 55:169–179
12. Han Q, Ding Z, Xueping X, Wang T, Chu F (2019) Stator current model for detecting rolling
bearing faults in induction motors using magnetic equivalent circuits. Mech Syst Signal Process
131:554–575
13. Nandi S, Lee SB (2011) Detection of eccentricity faults in induction machines based on
nameplate parameters. IEEE Trans Ind Electron 58(5)
14. Park Y, Choi H, Shin J Park J, Lee SB, Jo H (2020) Airgap flux-based detection and classification
of induction motor rotor and load defects during the starting transient. IEEE Trans Ind Electron
67:10075–10084
Intelligent Technique Proposed
for Nonlinear Inductor Modelling
for DC/DC Converters
Abstract The inductor is one of the main components for converting energy within
power converters. The inductance value does not always be constant as it decreases
with the current, particularly in the saturation region where the current’s value
increases. Thus, the inductor decrease affects the dynamic of the power converter.
To exploit its nonlinear behavior and to achieve a precise dynamic behavior of the
converter, an accurate inductor model must, therefore, be considered. Several models
has been dedicated to analyze these characteristics, but they are complicated and diffi-
cult to be implemented. In this paper, the Adaptive Neuro-Fuzzy Inference system
(ANFIS) model is used to solve the nonlinear behavior of the inductor. The proposed
ANFIS model is a clever neuro-fuzzy technique, which has proved to model and
control diverse processes that are nonlinear. The main feature of this model is that
the evolution of the flux vs current curve behavioral analysis has been taken into
account using an experimentally characterized set of data. The study is set up on the
DC/DC buck converter, using Matlab/Simulink. The results obtained by simulation
provide the inductor current and voltage prediction, which have good agreement with
the experimental measurements with relative error about 25%.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1387
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_126
1388 R. El idrissi et al.
To test the proposed model for the nonlinear inductor taking into account the satura-
tion, a buck converter, working in CCM, is used. The inductor is excited by a square
waveform provided by the converter. The power inductor of all switching converters
conducts DC and triangular ripple currents, in a similar manner. The buck converter
is also a good way to check and illustrate how saturating conditions usually meet
Intelligent Technique Proposed for Nonlinear Inductor Modelling … 1389
The relation between magnetic flux and current is widely considered to characterize
an inductor:
= Li (1)
The current vs voltage ratio is obtained by the Faraday law for the linear inductor,
which considers inductance to be constant:
d di
VL = =L (2)
dt dt
In most studies assume that the inductance L is constant. Indeed, in practice the
inductance varies with respect to current in the inductor. In particular, the inductor
core magnetic flux is a nonlinear function of the current as follow:
= (i) (3)
(a) (b)
Fig. 3 Flux value against current (a) Inductance value against current (b)
Intelligent Technique Proposed for Nonlinear Inductor Modelling … 1391
d d(i L ) d(i L ) di L di L
VL = = = = L(i L ) (4)
dt dt di L dt dt
Where:
∂(i L )
L(i L ) = (5)
∂i L
(FIS), an integrator to get the flux, and an enforced current source, which transforms
the input signal into a corresponding current source.
(a) (b)
(a) (b)
Fig. 6 ANFIS surface viewer (a) ANFIS rule viewer (b)
input parameter. The second column is the output variable’s membership function,
the inductor current. The rule Viewer simultaneously interprets the complete fuzzy
inference mechanism.
Figure 7 validates the precision of the current-flux curve modelling using ANFIS
model. It contrasts the real value of the inductor with the values expected by the
ANFIS model. It is obvious from Fig. 7 that the ANFIS predicted values well
represent the real values of the entire set of test data.
Fig. 7 Comparison of the inductor current values predicted by ANFIS models and the actual values
of data set and the error between them
1394 R. El idrissi et al.
5 Simulation Results
The DC/DC buck converter was introduced in Matlab/Simulink taking into account
the inductor’s parasite effects and not the use of a linear inductor; the proposed model
in Fig. 5 was taken into account. In this simulation, the load has been changed at
0.008 s and the frequency has been lowered to 6 kHz. Figures 9a, b and c show the
inductor current, the magnetic flux, and the variation of the inductance value in the
step response. The step response is obtained by changing the load resistance from
70 to 13 . It is obvious that when the current increases, the inductance declines
and the current ripples rise. Magnetic flux also increases according to the inductor
current.
Figure 9d shows the comparison of the proposed model with linear inductance
model at load change. After load changes, the current in the proposed model, for
the inductor core saturation effect, is rising at a much higher rate. Inductance value
significantly decreases. Hence, the current ripples rise. In contrast, the linear model
does not consider saturation. Figure 9d gives the difference between the nonlinear
and linear inductor model.
6 Experimental Results
The circuit was also hardware designed for testing the simulation efficiency. B82615-
B2302M EPCOS inductor is used. Figure 11 illustrates the experimental system photo
(Fig. 10).
The model is tested by comparing the current and the voltage of the inductor.
Figure 11 reflects the current form and voltage that passes through the inductor while
the frequency of the switch is high and low. In Fig. 11a, the switching frequency is
20 kHz with a supply voltage 30 V and a duty cycle equals to 50%. It can be observed
Intelligent Technique Proposed for Nonlinear Inductor Modelling … 1395
(a)
(b)
(c) (d)
Fig. 9 The buck converter’s simulation results based on the proposed inductor model
from this figure that the shape of the current is almost triangular. In Fig. 11b, the
switch frequency is decreased to 5 kHz and the voltage of supply is raised to 60 V
while the duty cycle is kept at 50%; it can be noted from the figure that the current
ripples are increased, the peaks on the current, and some fluctuations on the converter
output voltage.
Figure 12a displays the experimental and simulation current and Fig. 12b indicates
the relative error. It is observed that in the in the worst case, the loss is smaller than
15% in a switching period.
The simulation of the proposed method, the linear, and experimental waveforms
of the current are illustrated in Fig. 12c. Figure 12d shows the zoom of the last figure.
1396 R. El idrissi et al.
(a) (b)
Fig. 11 The buck converter assessed high frequency (a) and low frequency (b) experimental
waveforms: output voltage (blue), the inductor current (red)
(a) (b)
(c) (d)
It is observed in Fig. 12c that the inductor current of experimental circuit and the
proposed model are much closed and the difference is largest between the proposed
and the linear models. Therefore, the proposed model performs well to keep up the
current waveforms. Figure 13 shows waveforms of the inductor voltage obtained in
experimental and simulation studies.
Intelligent Technique Proposed for Nonlinear Inductor Modelling … 1397
7 Conclusion
A new analytical model, considering the nonlinear behaviour of inductor core, has
been proposed in order to exploit the inductor saturation. The model is based on
the ANFIS technique. The ANFIS model is trained using a separate collection of
measured data of magnetic flux vs inductor current. The proposed technique is char-
acterized by its simplicity and does not require a complicated mathematical equation.
In a Buck converter, the nonlinear model for the inductor has been implemented in
Matlab/Simulink and experimentally evaluated. The study of performances indicates
a very close coordination between inductor current ripple waveforms of the proposed
model and the experimental waveforms, which proves the proposed model prediction
accuracy.
References
1. Palma L, Todorovic MH, Enjeti P (2006) Design considerations for a fuel cell powered DC-DC
converter for portable applications. In: 21st IEEE APEC 2006, p 6
2. Vazquez MJV, Marquez JMA, Manzano FS (2008) A methodology for optimizing stand-alone
PV-system size using parallel connected DC/DC converters. IEEE Trans Ind Elect 7:2664
3. Fadil HEl, Giri F, Magueri OEl, Chaoui FZ (2009) Control of DC–DC power converters in the
presence of coil magnetic saturation. Cont Eng Pract 17:849
4. Salas RA, Pleite J (2011) Simulation of the saturation and air-gap effects in a POT ferrite core
with a 2-D finite element model. IEEE Trans Magn 47:4135
1398 R. El idrissi et al.
5. Salas RA, Pleite J (2013) Equivalent Electrical model of a ferrite core inductor excited by
a square waveform including saturation and power losses for circuit simulation. IEEE Trans
Magn 49:4257
6. Stadler A, Stolzke T, Gulden C (2012) Nonlinear power inductors for large current crest factors.
Proc PCIM 1548
7. Salas R, Pleite J, Olias E, Barrado A (2006) Nonlinear saturation modelling of magnetic
components for circuit simulation. IEEE. INTERMAG, 993
8. Gurleyen H, Erkan M (2017) Analytical modeling of magnetically saturated inductance by
lambert W function. J Magn 22:369
9. Lullo G, Scirè D, Vitale G (2017) Non-linear inductor modelling for a DC/DC buck converter.
Ren En Pow Qual J
10. You BG, Lee SW, Choi GB, Yoo DW, Lee BK (2011) Comparison of simulation and
experimental results for Mega Flux inductors in hybrid electric vehicles. In: 8th IPCE
11. You BG, Lee BK, Lee SW, Choi GB, Yoo DW (2011) Experimental comparison of mega-flux
and JNEX inductors in high power DC-DC converter of hybrid electric vehicles. IEEE. VPPC
1
12. Susitra D, Paramasivam S (2014) Non-linear flux linkage modeling of switched reluctance
machine using MVNLR and ANFIS. J Int Fuz Syst 26:759
13. Bassoui M, Ferfra M, Chraygane M (2016) ANFIS modeling of nonlinear inductance. In: 2nd
ICEIT
14. Jang JSR, Sun CT (1995) Neuro-fuzzy modeling and control. Proc IEEE 83:378
15. Jang JSR (1993) ANFIS: adaptive-network-based fuzzy Inference Systems. IEEE Trans Syst
23:665
High Order Sliding Mode Power Control
of Doubly Fed Induction Generator
Based on Wind Turbine System
Abstract This paper discusses the control of the active and reactive stator power of
a doubly fed induction generator (DFIG) based on a wind energy conversion system
using super-twisting sliding mode control (STSMC) control. The objective of this
work is to evaluate the robustness of STSMC control against the variation of internal
parameters of the DFIG, and also to reduce the power ripple resulting from the
chattering phenomenon which is the major drawback of the conventional SMC. The
modeling of the DFIG and the implementation of the proposed control strategy are
performed and tested using the numerical simulation environment Matlab/Simulink.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1399
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_127
1400 Y. Dbaghi et al.
controls such as backstepping and sliding mode control, and the intelligent controllers
based on fuzzy logic and neural networks.
Boubzizi, S. et al. have demonstrated in [2] that the PI controller is limited in
the linear functioning of the system otherwise it diverges. Kelkoul B. et al. have
concluded that the use of the conventional sliding mode offers acceptable robustness
and a fast response time except that it leads to an undesirable chattering phenomenon
[3].
In our paper, the objective is to design a high order sliding mode controller
that allows both the minimization of the chattering phenomenon and ensures high
robustness and rapid response against unpredictable internal disturbances of the
machine.
This article is organized as follows: in Sect. 2 we present the modeling of the wind
turbine. Section 3 will be devoted to the modeling and vector control of the DFIG.
The design of the proposed control is described in Sect. 4. The 5th section presents
the simulation results and their Interpretations.
The following formula presents the power produced by the wind turbine as [4]:
1
Paer = Cp (λ, β).ρ.π.R2 .V3 (1)
2
High Order Sliding Mode Power Control of Doubly Fed Induction Generator ... 1401
Where: ρ is the air density, R is the radius of the blades, V is the wind speed, β is
the pitch angle.
λ is the speed ratio defined by [5]:
R.t
λ= (2)
V
t : The angular speed in the shaft. of turbine.
Cp(λ, β) is the power coefficient represent the efficiency of the wind energy
conversion system, which given by [6]:
Cp (λ, β) = C1 (C2 × A − C3 β − C4 )e−c5 ×A + C6 λ
(3)
A = λ+0.08β
1
− 0.035
β3 +1
Where, mec is the mechanical speed of the generator, Tg is the generator torque,
and Tem is the electromagnetic torque.
Newton’s second law of motion on the generator shaft gives [8]:
Fig. 3 Block diagram model of wind turbine system with TSR MPPT algorithm
Tg − Tem − f = J ddtmec
(5)
J = GJt + Jg
With: J is the total inertia, Jg is the generator inertia and, f is the viscosity
coefficient, and Jt is the turbine inertia.
The MPPT algorithm used is based on TSR (Tip Speed Ratio).the optimal
electromagnetic torque expression for the maximum power is defined as [8] (Fig. 3):
The stator and rotor DFIG’s voltages in the park frame are expressed as [9]:
⎧
⎪
⎪ Vds = Rs Ids + d
ϕ
dt ds
− ωs ϕqs
⎨
Vqs = Rs Iqs + d
ϕ
dt qs
+ ωs ϕds
(7)
⎪
⎪ V = Rr Idr + d
ϕ − ωr ϕqr
⎩ dr dt dr
Vqr = Rr Iqr + d
ϕ
dt qr
+ ωr ϕdr
With: Rr and Rs are respectively the rotor and stator resistances per phase.
The stator and rotor flux of DFIG are defined by:
High Order Sliding Mode Power Control of Doubly Fed Induction Generator ... 1403
⎧
⎪
⎪ ϕds = L s Ids + M Idr
⎨
ϕqs = L s Iqs + M Iqr
(8)
⎪
⎪ ϕ = L r Idr + M Ids
⎩ dr
ϕqr = L r Iqr + M Iqs
Where: Idr , Iqr , Ids , Iqs , are respectively the rotor and stator currents in dq-axis. Lr ,
Ls are the rotor and stator inductances. M is the DFIG mutual inductance.
The expression of the stator active and reactive powers is defined by [10]:
Ps = Vds Ids + Vqs Iqs
(9)
Q s = Vqs Ids − Vds Iqs
pM
Tem = (Idr ϕqs − Iqr ϕds ) (10)
Ls
The value of the stator resistance is small as that can be neglected, then stator
voltages can be rewritten as:
Vds = 0
(12)
Vqs = Vs = ωs ϕs
From Eq. (8), Eq. (9) and Eq. (12), the stator active and reactive powers can be
written as function of rotor currents as:
Ps = Vs LMs Iqr
2 (13)
Qs = ωVssLs Vs LMs Idr
By combining the previous equations, the rotor voltages (d, q) in terms of the
rotor currents (d, q) are expressed as:
⎧
⎨ Vdr = Rr Idr + Lr − M2 dIdr + eq
Ls dt
(14)
⎩ Vqr = Rr Iqr + Lr − M2 dIqr + ed
Ls dt
⎧
⎨ eq = −gω Lr − M Iqr2
s L s
(15)
⎩ ed = gωs Lr − M2 Idr + gMVs
Ls Ls
1404 Y. Dbaghi et al.
Where: g is the slip, and ed , ed represent the coupling terms between the dq-rotor
voltages.
The sliding mode is one of the most powerful nonlinear control techniques that is
based on the stability of the Lyapunov. This technique was developed by V. I. Utkin in
[12], its main operating principle lies in its ability to force a system’s state variables to
slide in a specific area called the sliding surface around a desired state of equilibrium.
This is achieved while respecting stability in the sense of Lyapunov. Its weak point
is the phenomenon of chattering caused by the discontinuous part of the control, this
phenomenon degrades and causes undesirable functioning of the system. The use
of a super-twisting sliding mode control presented a solution to this problem by its
ability to reduce this phenomenon to minimum as possible.
The STSMC The control law is given by [12]:
This control law consist of a continuous control VC (t), and a super-twisting control
Vst (t) that is defined by [13]:
With:
√
v1 (t) = −α |S|.sat (S)
(18)
v2 (t) = −γ .sat (S)
By deriving Eq. (20) and using Eqs. (13), and (14) we get:
High Order Sliding Mode Power Control of Doubly Fed Induction Generator ... 1405
Ṡ(Ps ) = Ṗs∗ + σL
Vs M
s Lr
Vqr − Rr Iqr − ed
(21)
Ṡ(Qs ) = Q̇s∗ + σLVs M
s Lr
Vdr − Rr Idr − eq
M2
With: σ = 1 − Lr Ls
.
In the sliding mode we have S(Ps , Qs ) = Ṡ(Ps , Qs ) = 0. Then we can express the
continuous controls of the dq rotor voltages by:
Vqr (C) = −Ṗs∗ σLs Lr
Vs M
+ Rr Iqr + ed
∗ σLs Lr (22)
Vdr (C) = −Q̇s . Vs M + Rr Idr + eq
To ensure the stability of system control we should respect the Lyapunov condi-
tion S(Ps , Qs ) × Ṡ(Ps , Qs ) < 0, then we should have α1 , α2 , γ1 and γ2 as positive
constants.
The results of this section show the MPPT operation of the wind turbine based on
the DFIG. We applied a random wind profile (Fig. 4 (a)) of a mean value of 9 (m/s)
to the wind turbine. The results obtained show a successful operation while ensuring
an optimal tracking of the maximum power (Fig. 4 (b)). This is verified by the power
coefficient maintained at its maximum value according to Fig. 5.
Fig. 4 (a) Wind turbine profile (b) MPPT stator active power response
1406 Y. Dbaghi et al.
In order to test and validate the performance of our proposed control method, we have
attacked the rotor side converter of the DFIG with two profiles representing active
and reactive power references such as illustrated by the blue traces in Fig. 6. The
simulation results shown in Fig. 6 describe the responses of the active and reactive
stator powers of the generator under two tests, a normal operation and an operation
disturbed by an increase of the stator resistance and the stator inductance of the DFIG
by 50% of their nominal values.
The simulation results show a great tracking of the active and reactive power
references and the chattering phenomenon is almost negligible for the normal test
(red trace) and also for the robustness test (pink trace). We can also notice that the
response time is quite fast: tr5% = 1.5(ms) in the normal test. The robustness test
has only shown a degradation of the response time, which has increased to 5.2(ms)
that remains just as fast. It is also observed that the decoupling of the q-axis and
the d-axis has been perfectly achieved, allowing an independent control of the stator
active and reactive power.
Figure 7 shows the evolution of the stator and rotor currents of DFIG. It is clear
that they vary according to the stator powers references.
By comparing our results with others published in the literature, our approach
shows a very fast response time with a negligible chattering phenomenon in compar-
ison to the results obtained by the conventional SMC proposed by Y. Bekakra et al.
In [14]. For the backstepping control proposed by L. Saihi et al. in [15], our method
offers a perfect decoupling with respect to their controller, which presents a poor sepa-
ration between the dq-axis of Park, which have caused of intense and undesirable
peaks in the active and reactive power responses.
High Order Sliding Mode Power Control of Doubly Fed Induction Generator ... 1407
Fig. 6 (a) Stator active power response (b) Stator reactive power response
Fig. 7 (a) Stator currents response (b) Park frame rotor currents response
6 Conclusion
This paper presents the control of the active and reactive stator power of the DFIG by
Super-twisting SMC according to the DFIG model with stator flux orientation. The
simulation results have largely satisfied the required objectives, they have shown a
high robustness against variations of the machine’s internal parameters, a fast enough
response time, and the chattering phenomenon has now been nearly neglected. From
these results it can be concluded that the use of a high order sliding mode controller
can significantly enhance the performance of a DFIG based wind energy system.
Appendix
References
1. Tahir A, EL-Mukhtar M, EL-Faituri A, Mohamed F (2020) Grid connected wind energy system
through a back-to-back converter. Comput Electr Eng (85):106660
2. Boubzizi S, Abid H, El Hajjaji A, Chaabane M (2018) Comparative study of three types of
controllers for DFIG in wind energy conversion system. Prot Control Mod Power Syst 3
3. Kelkoul B, Boumediene A (2019) Nonlinear sliding mode control of DFIG based on wind
turbines. ICAIRES 2018, LNNS 62, pp 206–215
4. Mensou S, Essadki A, Nasser T, Bououlid Idrissi B, Ben TL (2020) Dspace DS1104 imple-
mentation of a robust nonlinear controller applied for DFIG driven by wind turbine. Renew
Energy 147:1759–1771
5. Nahilia H, Boudour M, Cardenas A, Agbossou K, Doumbia M (2019) Doubly fed induction
wind generators model and field orientation vector control design and implementation on
FPGA. Int J Dyn Control. 7:1005–1014
6. Farhat S, Alaoui R, Kahaji A, Bouhouch L (2018) Wind turbine MPPT strategy with DFIG
vector control. Int Rev Model Simul 11(6):406
7. Sabiri Z, Machkour N, Rabbah N, Nahid M, Kheddioui E (2017) Command of a doubly-fed
induction generator with a backstepping controller for a wind turbine application. Int Rev
Autom Control 10(1):56–62
8. Dbaghi Y, Farhat S, Mediouni M (2020) First order sliding mode and super-twisting sliding
mode control of doubly fed induction generator driven by wind turbine system: a comparative
study. J Adv Res Dyn Control Syst 12(04):1656–1667
9. Moradi H, Vossoughi G (2015) Robust control of the variable speed wind turbines in the
presence of uncertainties: a comparison between H∞ and PID controllers. Energy 90:1508–
1521
10. El Azzaoui M, Mahmoudi H (2017) Fuzzy control of a doubly fed induction generator based
wind power system : Int. J. Wseas Trans Power Syst (12)
11. Kiruthiga B (2015) Implementation of first order sliding mode control of active and reactive
power for DFIG based wind turbine. Int J Inf Futur Res 2(8):2487–2497
12. Utkin VI (1973) Sliding modes in multidimensional systems with variable structure. In Confer-
ence on Decision and Control including the 12th Symposium on Adaptive Processes, pp
727–727, IEEE ,San Diego, CA, USA
13. Han Y, Ma R, Pan W, Wang C (2019) A novel super-twisting algorithm-based direct power
control strategy for doubly fed induction generator. In 12th Asian Control Conference (ASCC),
pp 1619–1624, Kitakyushu-shi, Japan
14. Chalanga A, Kamal S, Fridman LM, Bandyopadhyay B, Moreno JA (2016) Implementa-
tion of super-twisting control: super-twisting and higher order sliding-mode observer-based
approaches. IEEE Trans Ind Electron 63(6):3677–3685
15. Bekakra Y, Ben Attous D, Bennadji H (2019) Sliding mode control of DFIG driven by wind
turbine with SVM inverter. In Hatti M. (eds) Renewable Energy for Smart and Sustainable
Cities. ICAIRES 2018. LNNS, (62). Springer, Cham
16. Saihi L, Bakou Y, Harrouz A, Colak I, Kayisli K, Bayindir R (2019) A comparative study
between robust control sliding mode and backstepping of a dfig integrated to wind power
system. In 7th International Conference on Smart Grid (icSmartGrid), IEEE
Robust Voltage Control for Four-Phase
Interleaved DC-DC Boost Converter
for Electric Vehicle Application
Abstract The electric vehicle and the quality of power electric are currently the most
requested areas of research in power electronics. In this context this paper presents
a robust control strategy for 4 phase interleaved bidirectional DC/DC for HEV/HV
application. The interlaced topology is used to reduce the current and voltage ripples
and to limit in current rating the inductor switch. Indeed, this paper aims to improve
the dynamics of the output voltage of this converter, two control strategies are
analyzed and conformed by simulation and experimental work. Advanced control
technique based on RST controller and PI controller. The work presented in this
article shows the advantage of the RST controller method over the PI controller, the
measurements correspond well to the simulated results.
1 Introduction
The Electric Vehicles (EVs) have advantages over existing cars until now, these cars
contribute to reducing CO2 emissions and earth warming and improving public health
and reducing ecological damage, which slows down climate change. The advent of
electric cars has reduced the intensive exploitation of natural resources as petroleum
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1409
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_128
1410 M. Barara et al.
and gas. EVs have turned to other alternative sources of electrical energy which has
improved the environment.
The EVs are essentially based on the bidirectional DC-DC converter which
connects the inverter to the energy storage system. During, the buck mode the energy
is sent back to the energy storage system, while in boost mode power is supplied
from the energy storage system to the drive system [1] as shown in Fig. 1.
In the EV system, the input voltage of the DC-DC converter boost is the battery
voltage and the current flowing through this system can flow in either direction
depending on the mode of operation of the vehicle. This is the bidirectional converter.
The boost converter is used in EVs to maintain the motor current at the desired
value in order to ensure motor voltage regulation. Thus it allows the regulation of the
speed of the engine and the torque and consequently of the vehicle in traction and also
in electric braking. The bidirectional converter is a DC-DC converter that allows us to
obtain from a fixed DC voltage source a DC voltage source with an adjustable value
of upper or lower value (boost or buck of voltage). It essentially consists of power
switches (transistors, diodes…) and passive components (inductors, capacitors …).
The DC-DC converter provides power to the vehicle’s onboard network (lighting,
computers, air conditioning…) and recharges the low voltage battery from the high
voltage battery.
Among the topologies of the DC-DC converters existing in the literature, the
choice was based on the interlacing topology because it is very suitable for the system
studied in this research work given the minimization of the nominal current of the
inductance switch and reduction of current and voltage ripple, smaller component
size, improved output impedance properties. Several control techniques have been
proposed in the literature to control converters in order to solve the problem of the
load variation. Among the classic converter control techniques there is the PI, PID,
fuzzy logic controller or a mixed controller between them [1].
The interleaving technique allowed to reduce input/output voltage ripples, and
minimize power loss to provide higher efficiency as well as reduce the weight and
size of passive converter components while reduction can also be performed by
increasing the switching frequency.
In reference [2], the mode of operation of the converter with the input current
ripple lowest compared to regular inductance, and then determine the appropriate
Robust Voltage Control for Four-Phase Interleaved DC-DC Boost Converter ... 1411
range of duty cycle. As the reference [3] proposes for the control of the DC-DC
converter the controller in sliding mode in order to reduce the ripples in the output
voltage of the converter and the input currents. Finally, reference [4] also proposes
a digital control based on the synchronization and the phase shift operation. While
this paper focuses on improving the dynamics of the output voltage of a DC-DC
converter, within this framework two control strategies are analyzed in simulation
and tested on an experimental step for one phase.
In order to improve the regulation of the output voltage of the DC-DC converter
influenced by any disturbance, we propose an advanced control technique based on
RST controller which has proved its robustness in terms of stability, response fast and
dynamic and also in terms of simplicity to implement. The RST controller is tested
experimentally with a single-arm DC-DC converter using the dSPACE controller;
the results once again confirm the robustness of the control technique proposed in
this work.
PI controller coefficients are designed to compare the behavior of this classic
controller with advanced RST controller. The simulation results showed that the
RST controller gave transient rapid response with less startup undulations stable
than the PI controller.
This paper is organized in 5 sections. The Sect. 2 is dedicated of the studied system
description. The advanced RST controller is presented in the Sect 3. In the Sect 4 the
design of the RST controller is presented. The last section is devoted to the results
and discussions.
Where Yr (k) is the reference input. The following figure shows the block diagram
of the RST controller [7] (Fig. 3).
3 Simulation Results
The simulation work has been carried using Matlab/Simulink. The model of the
studied system presented in Fig. 1 is simulated by applying the parameters of the
Table 1. In this section, the studied system has been simulated using a four arms
converter.
Figure 2 shows the control scheme that has been designed in the case where the
DC-DC converter is powered by an ideal source on the input side, and a variable
load on the output side. This scheme has been tested in several configurations. The
parameters of the converter support are given in Table 1.
We varied the load (between 0 and 0.8 s and from 0.8 s to 1.4 s and the reference
voltage between 1.4 s and 1.8 s) to see the response and the behavior of the studied
system especially when we have applying two controls type on this system.
Figure 4 shows the output voltage of the Four-phase converter with PI controller.
3.1.1 PI Controller
A smalls peaks in the output voltage is observed in Fig. 4 due to sudden variation
of load and reference voltage, but it recovers quickly by the corrective action of
the PI controller. It is observed that the value of the output voltage is maintained
Table 1 Simulation
Parameter Value
parameters
Input voltage 200 V
L1,L2,L3 and L4 2 mH
C 26000 µF
R1 32
R2 16
F 20 kH
1414 M. Barara et al.
Now, the RST controller is applied. We can see the improvement carried by the
advanced controller by reducing the ripples of the output voltage also at the level of
the pursuit of the reference voltage.
Figure 5 shows the output voltage of the system studied after applying the
advanced control. The output voltage precisely pursues its reference without ripples
when the load is varied (between 0 and 0.8 s; from 0.8 s and 1.4 s and reference voltage
between 1.4 s and 1.8 s), which proves the efficiency of the proposed controller.
Finally, it’s clear that even if varied when the reference voltage and load the
output voltage perfectly follows its reference. This result proves the effectiveness of
the proposed method of control in the simulations works.
4 Experimental Results
In this section we present the results of the experimental setup. We limited the
experimental setup to one arm of the converter using a dSPACE card.
Figure 6 shows the experimental setup of the studied system with the different
devices used in work.
Figure 7 show the experimental performance of the control scheme for the output
voltage under variable reference voltage to prove the efficiency of the proposed
control method, by applying the classic PI controller.
4.1.1 PI Controller
We varied the reference voltage while the PI controller is applying, it s clear that the
output voltage follows its reference with a fairly long delay time (Fig. 7).
This time we applied the advanced RST controller, it is obvious that the output voltage
follows precisely its reference with a faster response time, a reduced overshoot
(Fig. 8).
(a) Output voltage under sudden variation of reference voltage with PI controller
(b) Zoom of Output voltage under sudden variation of reference voltage with PI controller
4.2.1 PI Controller
After varying the load, we can clearly see that we observe ripples and small peaks
(Fig. 9).
This figure shows that the experimental work confirms the simulation results, after
applying the advanced RST controller and even if when the load is changed the value
of the output voltage remains stable and this voltage follows its reference perfectly
without ripple and small peaks (Fig. 10).
Robust Voltage Control for Four-Phase Interleaved DC-DC Boost Converter ... 1417
(a) Output voltage under sudden variation of reference voltage with advanced controller
(b) Zoom of output voltage under sudden variation of reference voltage advanced controller
5 Conclusion
The study portrayed in this paper purposes to improve the quality both output of
current and voltage of the DC-DC converter connecter to an electric vehicle. To
overcome this problem, a boost converter was adopted introducing a new controller
design for regulating a DC bus voltage. The proposed design of the controller is based
on RST controller. Theoretical study of the RST approach was dressed and verified
through computer simulations performed under MATLAB/Simulink software. The
simulation results and experimental works are organized in subsections. The objective
was to prove the effectiveness of the RST controller when compared to PI controller.
The comparative study and analysis are demonstrated the superiority of the RST
controller which it s presents a good regulation of voltage, a faster response time, a
reduced overshoot, and a lower static error.
1418 M. Barara et al.
(b) Zoom of output voltage under sudden variation of load with PI controller
Fig. 9 Output voltage with PI controller
Robust Voltage Control for Four-Phase Interleaved DC-DC Boost Converter ... 1419
(a) Output voltage under sudden variation of load with advanced controller
(b) Zoom of output voltage under sudden variation of load with advanced controller
References
1. Schumacher D, Magne P, Preindl M, Bilgin B, Emadi A (2016) Closed loop control of a six
phase interleaved bidirectional DC-DC boost converter for an EV/HEV application. In IEEE
Transportation Electrification Conference and Expo (ITEC), pp 1–7. Dearborn
2. Kascak S, Prazenica M, Jarabicova M, Konarik R (2018) Four phase interleaved boost converter-
analysis and verification. Acta Electrotechnica et Informatica 18(1)
3. Ayoubi Y, Elsied M, Oukaour A, Chaoui H, Slamani Y, Gualous H (2016) Four-phase interleaved
DC-DC boost converter interfaces for super-capacitors in electric vehicle application based on
advanced sliding mode control design. Electric Power Syst Res 134:186–196
4. Nagulapati K, Rangavalli V, Vanajakshi B (2014) Modelling, analysis and simulation of 4- phase
boost converter. Int J Electr Comput Energ Electron Commun Eng 8(9)
5. Nisak K, Aida SA, Zainul AZ (2018) Interleaved 4-Phases DC-DC converter for electric
scooter to improve efficiency. In: MATEC Web Conference Volume 225, 2018 UTP-UMP-VIT
Symposium on Energy Systems
1420 M. Barara et al.
Abstract The wind turbine coupled with a doubly fed induction generator (DFIG)
demonstrates good robustness and energy performances against the intermittent
nature of the wind, which allows many control strategies driving this system to
control it in the right conditions. Besides, the wind energy is completely depending
on the wind flow which is variable and can overpass the rated value of the wind speed.
In this case, the wind turbine should be sheltered versus load stress and the poten-
tial risk of destruction. In this paper, to profit from wide wind speed and exploiting
the totality of wind energy, an artificial neural network is designed to implement a
maximum power point tracking MPPT-pitch angle control strategy for controlling the
wind turbine. This strategy lets the DFIG extract the maximum power and protects
the wind turbine system from the overload of the electromagnetic torque when the
wind speed is higher than its rated value. The simulation results are performed using
the Matlab/Simulink environment and show well control performances under a wide
wind speed range.
Keywords Wind turbine · DFIG · Neural network · MPPT · Pitch angle control
1 Introduction
Nowadays, to face the challenges of increased energy demand and climate changes,
several renewable energy sources are committed to meeting the expectations of manu-
facturers and consumers. Among these renewable energy sources, wind energy is
considered a vital energy source that relies on the kinetic energy extracted from the
wind.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1421
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_129
1422 M. Reddak et al.
Due to the intermittent nature of the wind, new wind turbines are equipped
with a robust energy conversion system [1, 2] for achieving maximum energy effi-
ciency. Among the components of this system, a generator is used for converting
mechanical energy to electrical one. Currently, (DFIG) [3, 4] is extensively used
as an electromechanical structure in wind energy conversion systems because of its
advantages.
To benefit of full range of the wind speed, control strategies are designed to reach
the maximum power once the wind speed is average, and also to preserve the power
of the wind turbine at its maximale value while the wind speed is upper relying on
the pitch control [5].
In this context, several robust control strategies, for example, nonlinear controls
(sliding mode, backstepping) [6] and even fuzzy logic control [7], are used to drive
the wind turbine in different wind regions. Artificial neural network (ANN) control
also meets this need since it has given good results in the areas of complex problem
solving and robust control [8–10].
The objective of this paper is the control of the wind turbine in large wind speed
by adopting the pitch control. The pitch angle is generated by a neural network while
the wind speed is upper to the limit value. Also the system control strategy aims to
warrant the taking out of the maximum power once the wind speed is less.
This paper is planned as follow: in Sect. 2, the model of the system is developed.
In Sect. 3, the pitch angle control using neural network is described. The strategy
control of the system will be introduced in Sect. 4 and before conclusion, some
simulation results will be discussed in Sect. 5.
The wind turbine equipped with the DFIG system, shown in Fig. 1, contains three
elements: the generator (DFIG), the turbine and the power converters. These later are
divided into two converters, a rotor side converters (RSC) which controls the powers
delivered by the generator, and a grid side converter that controls the DC-bus. The
control of the grid side converter is not concerned in this paper.
The power taken from kinetic energy, according to Betz’s theory, is defined as
follows:
1
Pa = ρπ R 2 v 3 C p (λ, β) (2)
2
According to this theory, the extraction of the power produced by the wind flow
through the wind turbine cannot be fully achieved. But this power is fractionated by
the coincidence effect of the wind with the blades. To define this limit, the power
coefficient C p that depends on this aerodynamic power is defined in Eq. 3:
150.9 − 18.4
C p (λ, β) = 0.46( − 0.579β − 0.00201β 2.14 − 13.21)(e λi ) (3)
λi
λi is expressed in Eq. 4:
1 1
λi = + 3 (4)
λ + 0.002β β +1
Rt
λ= (5)
v
Figure 2 gives the curve of C p of different pitch angle values.
And Eq. 6 gives the obtained aerodynamic torque:
1 Cp
Ta = ρπR3 v 2 (6)
2 λ
1424 M. Reddak et al.
With simplified notation and by relating the Park reference to the stator field, the
expression of the DFIG voltages and fluxes in dq reference are given by Eq. 7.
⎧
⎪
⎪ v = Rs ids + dϕdtds − ωs ϕqs
⎪ ds
⎨ dϕ
vqs = Rs iqs + dtqs + ωs ϕds
(7)
⎪
⎪ vdr = Rr idr + dϕdtdr − ωr ϕqr
⎪
⎩ v = R i + dϕqr + ω ϕ
qr r qr dt r dr
ϕds = L s i ds + Mi dr = ϕs
(8)
ϕqs = L s i qs + Mi qr = 0
M
Tem = − pϕs irq = K T irq (11)
Ls
3 Neural Network
nl
y lj = f ( wlji xi + bil ) (13)
i
With
xi : Outputs of the previous neurons.
wlji : The weight of synapses binding neurons.
bil : Coefficient for adjusting the neural network.
The function f can be presented by hyperbolic, threshold, sigmoid or radial
functions.
This ensemble of interconnected formal neurons allows the resolution of complex
problems by adjusting the weight coefficients using a learning algorithm.
In our case, the ANN bloc, that control the pitch angle, is a static multi-layer
perceptron MLP. The architecture of this later is revealed in Fig. 3.
According to Fig. 3, the neural network takes as input the ratio of the tip speed λ
and C p , and as output the pitch angle β.
The performance training of the proposed neural network is depicted in Fig. 4.
Figure 5 illustrates the curve of the wind turbine power across the wind speed; there
are three operation modes: In the first region, before the cut-in speed, there is no
production of the power. Then in second region, that matches to the average wind
speed, it is necessary to control the DFIG to benefit from the maximum power. In
the last region, where the wind speed exceeds the passage value which is 12 m/s, it
is necessary to ensure the control of the pitch angle to conserve the rated power.
1426 M. Reddak et al.
Fig. 3 The proposed MLP neural network for pitch angle control
Figure 6 shows the wind turbine coupled with the DFIG system and the used
control strategy. The control system is separated to two parts: the control at the
turbine side and the control at the DFIG side.
The first part maintains the power captured at its nominal value by acting on the
pitch angle control in region 3, it also ensures the speed control of the generator
according to the measured speed of the wind by optimizing the power extracted in
region 2 regarding this equation.
MPPT and Pitch Angle Based on Neural Network Control of Wind Turbine Equipped ... 1427
Fig. 6 Control strategy of the wind turbine driven a DFIG using an ANN
1
Popt = ρπ R 2 v 3 C p_max (λopt , βmin ) (14)
2
Therefore, this part gives the reference torque necessary for the active power
control by generating the rotor current reference irq r e f .
On the side of DFIG control, the converter of the machine-side controls the
generated active and reactive powers.
1428 M. Reddak et al.
5 Simulation Results
To confirm the control strategy approach of the wind turbine equipped with a DFIG,
a wind profile, described in Eq. 1, has been adopted and potted as depicted in Fig. 7.
Indeed, it can be verified, in Fig. 8, that the pitch angle produced by the artificial
neural network, varies according to this wind profile. It is at minimum value (0°)
when the wind speed is under the transition value (12 m/s). Whereas in region 3,
which reflects the wind speed exceeding the pass through value, the pitch angle
takes the corresponding value other than the minimum one. With these strategies,
the MPPT and pitch control has been achieved.
Also it can be shown in Fig. 9 that the Cp is maintaining in the maximum value,
that is around 0.45, when the wind speed reflects the region 2. Although, the Cp is less
when the wind speed surpasses the transition value, which endorse the robustness of
the control under neural network.
The electromagnetic torque shown in Fig. 10 is limited to protect the system from
damaging and load fatigue while the wind speed is upper to the pass trough value.
Besides, as depicted in Fig. 11, the stator power is limited while the wind speed
is higher than the pass through value to give the nominal value that is 2e6 W, and
when the wind speed reveals the region 2, the regulator forces the DFIG to extract
the maximum power.
6 Conclusion
An intelligent ANN control, which is used for the MPPT and the pitch angle strategy,
has been suggested in this paper, to profit from the power delivered by a wind turbine
equipped by DFIG under full range of wind speed. For the purpose of protecting
the wind turbine from the fatigue load when the speed surpasses the rated value, the
electromagnetic torque has been limited. In this case, the pitch angle, generating using
an ANN control block, and power coefficient are set by producing a corresponding
value. However, when the wind speed is under the pass through value, and the pitch
angle given by ANN is at a minimum value to extract minimum value to ensure
maximum power extracting. The control strategy potency using ANN of wind turbine
equipped by DFIG has been shown in the simulation results under Matlab/Simulink
environment .
Appendix
(Table 1)
References
1. Tripathi SM, Tiwari AN, et Singh D (2016) Optimum design of proportional-integral controllers
in grid-integrated PMSG-based wind energy conversion system: optimum design of PI
controllers. Int Trans Electr Energy Syst 26(5):1006−1031
2. Singh M, Khadkikar V, Chandra A (2011) Grid synchronisation with harmonics and reactive
power compensation capability of a permanent magnet synchronous generator-based variable
speed wind energy conversion system. IET Power Electron 4(1):122. https://doi.org/10.1049/
iet-pel.2009.0132
3. Bekakra Y, Attous DB (2014) Optimal tuning of PI controller using PSO optimization for
indirect power control for DFIG based wind turbine with MPPT. Int J Syst Assur Eng Manag
5(3):219−229
4. Boubzizi S, Abid H, El hajjaji A, Chaabane M (2018) Comparative study of three types of
controllers for DFIG in wind energy conversion system. Prot Control Mod Power Syst 3(1):21
5. Reddak M, Berdai A, Nouaiti A, Vlasenko V (2018) Collaboration of nonlinear control strategy
and pitch angle control of DFIG equipped wind turbine during all operating regions. Int J
Comput Appl 179(25):16−21
6. Reddak M, Berdai A, Gourma A, Boukherouaa J, Belfiqih A (2016) Enhanced sliding mode
MPPT and power control for wind turbine systems driven DFIG (doubly-fed induction
generator). Int Rev Autom Control IREACO 9(4):207
7. Karimi-Davijani H, Sheikholeslami A, Livani H, Karimi-Davijani M (2009) Fuzzy logic control
of doubly fed induction generator wind turbine. p 11
8. Zhang Z-Y, Wang K-S (2014) Wind turbine fault detection based on SCADA data analysis
using ANN. Adv Manuf 2(1):70−78
9. Fazelpour F, Tarashkar N, Rosen MA (2016) Short-term wind speed forecasting using artificial
neural networks for Tehran. Iran Int J Energy Environ Eng 7(4):377−390
10. Marugán AP, Márquez FPG, Perez JMP, Ruiz-Hernández D (2018) A survey of artificial neural
network in wind energy systems. Appl Energy 228:1822−1836
11. Abdelli A, Sareni B, Roboam X (2007) Optimization of a small passive wind turbine generator
with multiobjective genetic algorithms. Int J Appl Electromagn Mech 26(3−4):175−182
12. Mirecki A, Roboam X, Richardeau F (2007) Architecture complexity and energy efficiency of
small wind turbines. IEEE Trans Ind Electron 54(1):660−670
13. Abad G, Iwanski G (2014) Properties and control of a doubly fed induction machine. In Power
electronics for renewable energy systems, transportation and industrial applications, H. Abu-
Rub, M. Malinowski, et K. Al-Haddad, Éd. Chichester, UK: John Wiley & Sons, Ltd, pp
270−318
Cairo Metro Power Analysis
and Harmonic Mitigation
Abstract In Egypt, Traction became one of the most important transportation means
so many electrical projects related to the metro have been established. Metro electrical
network is supplied by 12-pulse converter rectifier substations. These rectifier stations
are of course nonlinear loads which result in several orders of harmonics to the
electrical grid. Moreover, some problems occurred due to the negative effect of
harmonics, for example, cables and transformers insulation failure due to fatigue
resulted from the high frequency voltage harmonic content.
This paper proposes a passive filter for this harmonic reduction. IEEE519-2014
recommendation sets the limits of current harmonics, and the resulted voltage distor-
tion at the point of common coupling. This work uses the measured current and
voltage for estimating the significant harmonic orders. A suitable filter can be
designed, and implement according to these measured harmonics. The effect of trans-
former impedance and changing the configuration of motor traction from DC series
motor into AC induction motor with speed drive is illustrated in the analysis section.
To overcome the harmonic distortion problem and to define the proper mitigation
method, recommendations have been concluded.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1433
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_130
1434 A. A. Elsadek et al.
1 Introduction
2 System Configurations
Cairo metro network is composed of two lines, line 1 and line 2, two high voltage
stations (HVS) that feed the two lines. HVS of Tura El Ballad 66/20 kV that supplies
17 rectifier stations in line 1. The second HVS is at Ramses 220 kV /20 kV which
feeds the rest of line 1 rectifier stations, and all rectifier stations of line 2.
In line 1, RS consists of three parts, two rectifier groups, each group has one power
transformer with voltage 20 kV/1228 v and 2.6 MW connected ϒ/ − ϒ to achieve
phase shift of 30° to cancel 5th and 7th harmonic order from source current. The
third transformer is an auxiliary for feeding the rectifier lighting loads. The output
DC voltage is 1500 V.
On the other hand, the RS’s in line 2 have the same configuration, however, it
contains one traction group, and the output dc voltage is 750 vdc. For line 1, there
are two types of traction motors; the first one is induction motor which was working
since 2015 and the second oldest one is DC series motors. Figure 1 describes the
rectifier station configuration which represents the building block of metro electrical.
The measurements were carried out using the power quality analyzer, Hioki 3196
used for recording source current, and voltage waveforms, harmonic content, power
factor, and other power quality indices which are required for harmonic assessment.
All data are collected and analyzed using statistical analysis to consider the full
period of measurements in terms of 99th percentile and maximum values according
to the recommendation of IEEE 519-2014 [5].
The first location of measurement is point “A” considering one group of traction
loads. The second measurement has been taken aFt location “B” to consider the two
Cairo Metro Power Analysis and Harmonic Mitigation 1435
groups of the substation and the auxiliary transformer which supply the substation
loads (Lighting, air condition, etc.).
The third measurement location is at point “C” which considers the effect of the
cable length. On the other hand, the last measurement location is at point “D” which
considers groups of rectifier substations.
Measurements of point “B” are indicated in the following figures as a sample.
Figure 2 indicates source voltage and current waveforms. The instantaneous value
of voltage THDv and THDi versus time for 24 h period compared to the predefined
limit are described in Fig. 3, and Fig. 4 respectively showing actual value exceeds
the standard limit.
4 Measurement Summary
Table 1 illustrates a summary of the power quality indices measured for THDv, and
THDi at different locations of points A, B, C, and D.
• It is noticed that the distortion rate increases in case of an increase of load current,
conversely at the no-load condition the voltage distortion is under the critical
1436 A. A. Elsadek et al.
limit, a combination between the non-load period and the start of moving the first
train indicates the difference in Fig. 5.
• The voltage individuals of orders 11th , 13th appeared at all measured points
as a result of the behavior of 12-pulse rectifiers in load condition, it’s called
characteristic harmonics.
• Despite of these results, an old study related to measurement analysis attempt
only in line 1 of the metro network in 2010. It is showed that the current distortion
is slightly out of specs, and consequently, it doesn’t have an effect on voltage
distortion to exceed the limits. By comparing both analyses, resulted that the
reason for increasing the harmonic distortion is related to adding new trains in
2015 to the operating system. These trains have induction motors driven by PWM
drives, while the oldest trains have dc series motors controlled by resistances, also
1438 A. A. Elsadek et al.
changing impedance, power of the new AC trains are important factors for high
harmonic value results.
• The distortion of line two is much more than line one, this is because of increasing
the value of the impedance of the traction transformer of line two than line one
[6], moreover, all trains are AC induction machine, and driven by PWM inverters.
• Common feeding of line 1 and line 2 through the point (D) has major effecting
on each other through a distorted voltage supplied. Up till modernizing line 1 it
is recommended to separate loads of line 2 on HVS Ramses from loads of line 1
to decrease the voltage distortion effect, moreover design proper filter on AC side
to line 2.
• Auxiliary transformers of 20 kv/400 v, voltage transformers (VT), and current
transformers (CT) are affected rapidly more than traction transformers 20 kv
/1228 v, which may be due to different values of K-factor.
• Slightly impact of auxiliary transformer this is due to the light consumed loads,
although having significant non-linear loads such as (led lamps, chargers and
inverters …), this result got by connecting and disconnecting the auxiliary
transformer on service at operating period and after the non-traction period.
• Reference to system configuration, it’s noticed that the resonant dc filter has a
slight effect as the measurements on AC side, maybe it has an effective rule on
dc side, this result was according to measuring at some (RSs) that have a filter on
dc side and the others haven’t a filter.
• It is noted that sometimes the transformers overloaded with about 1.5 of its
primary rate, due to changing some of the old trains with higher power air-
conditioned trains without increasing the power of the rectifier stations, and
headway increasing.
• The harmonic currents associated with high losses and temperature will decrease
the expected lifetime of transformers. Recommendation of resizing substation
supplied power transformer depends on many factors included in reference [7].
• In 2019, a new RS was put in service according to the extension plan of line 1;
the dc filter needs to be checked according to proper design for new trains.
Cairo Metro Power Analysis and Harmonic Mitigation 1439
As per the violation identified by the recorded current THDI , the voltage at the point
of common coupling THDv, distortion factor, and total power factor, the need to
mitigate the harmonics is arising. Few of the most prevailing methods used today
1440 A. A. Elsadek et al.
A passive filter is combined of capacitors and inductors that are tuned to resonate at
a single frequency, or through a band of frequencies.
Passive filters work by exhibiting different impedance values at the resonant
frequency. A filter connected in series should present high impedance to the harmonic
frequency that needs to be blocked. Although a series configuration is possible, it
is more common to connect filters in parallel. Such a shunt configuration diverts
harmonic currents to ground, and simultaneously provide reactive power, which
may be used to correct the power factor. As such, passive shunt filters are designed
to be capacitive at the fundamental frequency.
Passive filter design is conducted to mitigate 11th and 13th harmonic order [4, 14,
15]. According to IEEE design recommendation, filter design will be 10.5th instead
of 11th order, and 12.5th instead of 13th order [16].
V2 (KV)
Xeff = (2)
Q rate of filter(MVAR)
(20)2
Xeff = = 31746.03
0.0126
XC = h2 / h2 − 1 ∗ Xeff (3)
10.52
XC = ∗ 31746.03 = 32036.6
10.52 − 1
Cairo Metro Power Analysis and Harmonic Mitigation 1441
1
C11 = (4)
2Π f XC
1
C11 = = 0.099 μF
2π ∗ 50 ∗ 32036.6
Where Xeff : effective reactance of harmonic filter
Qeff : effective reactive power of harmonic filter, this value from the measure-
ment
VLL : System line to line voltage
XC : Capacitive reactance of harmonic filter capacitor at f tuned
XL : Capacitive reactance of harmonic filter reactor at f tuned
h: harmonic number.
Xc 32036.6
XL = = = 290.58 (5)
h2 (10.5)2
XL 290
L11 = = = 924 mH (6)
2πF 2π ∗ 50
The resistance of filter value (R) is calculated according to the following equation
by assuming the quality factor Q is 50 as in [11]
XL
R = (7)
Q
290
R= = 5.8
50
• Some individual currents exceed the limits; moreover, it does not affect its voltage
distortion, so there is no need for a high pass filter to eliminate from 23th to 50th
as the most significant effect of current is at 11th and 13th as in Eqs. (8), and (9).
( (I2)2 +(I3)2 + (I4)2 + . . . . . . . . . . . . In2 .
THDi = (8)
I1
1442 A. A. Elsadek et al.
THDi at point (B) is 78% and by eliminating the value of 11th and 13th of individual
current, the new value of THDi = 19% as considering filter compensation
( (V2)2 +(V3)2 + (V4)2 + . . . . . . . . . . . . Vn2 .
THDv = (9)
V1
THDv at point (B) is 8.2% and by eliminating the value of 11th and 13th
of individual voltage, the new value of THDv = .07% as considering filter
compensation.
• The optimum filter location and best area space is at point (B) of the incoming of
the rectifier substation.
6 Conclusion
7 Future Work
Reference to the previous argument, the recommended solution for metro harmonic
mitigation is the active filter. The active filter has the advantage of not only mitigate
the harmonics but also compensates the reactive power to improve both distortion
factor and displacement power factor. The active filter becomes a common industrial
solution; however, the design of it is a critical job. The location and the rating of
the active filter should be carefully selected to achieve the mitigation target to meet
the international standard with better harmonics handling performance and overall
system efficiency. This will be discussed in future work as a part of the metro continual
improvement strategy.
Cairo Metro Power Analysis and Harmonic Mitigation 1443
References
1. Smith JC, Hensley G, Ray L (2009) IEEE recommended practice for monitoring electric power
quality. IEEE Std 1159–2009 (Revision of IEEE Std 1159–1995), pp c1–81
2. Gunavardhini N, Chandrasekaran M, Thirumoorthy A (2012) A study on harmonics in Indian
railway traction. IOSR Journal of Electrical and Electronics Engineering 2:1–4
3. Navaneethakrishnan G, Muthial C (2017) Power quality in railway traction and compensation
by combining shunt hybrid filter and TCR. Automatika 57:610–616
4. Popescu M, Bitoleanu A, Dobriceanu M (2008) Harmonic current reduction in railway systems.
WSEAS Transactions on Systems 7:689–698
5. Sabin DD (2002) Analysis of harmonic measurement data. IEEE power engineering society
summer meeting 2:941–945
6. Nhavira S, Polimac V (2007) Quality of supply issues arising from DC traction loads on a
metro system
7. EN 50329 (2003) Railway applications – Fixed installations – Traction transformers
8. Lukic L, Djapic M, Lukic D, Petrovic A (2001) Aspects of design of power transformers for
noise reduction. Parameters 60076:10
9. Elhenawy, AAE-M., Sayed, MM, Gilany MI (2018) Harmonic cancellation in residential build-
ings. In: 2018 twentieth international middle east power systems conference (MEPCON), pp
346–351
10. Zhongming Y, Pong M, Lo W, Yuen K (1999) Harmonic evaluation of traction system by Monte
Carlo simulation. In: 30th annual IEEE power electronics specialists conference. Record (Cat.
No. 99CH36321), pp 424–429
11. Monem O (2019) Harmonic mitigation for power rectifier using passive filter combination. In:
IOP conference series: materials science and engineering, p 012013
12. Fathima F, Karthikeyan SP (2016) Harmonic analysis on various traction transformers in co-
phase traction system. Ain Shams Eng J 7:627–638
13. Persson J (2014) Comparing harmonics mitigation techniques. In: Comsys AB, Lund, p 6
14. Vasanthi V, Ashok S (2011) Harmonic filter for electric traction system. In: ISGT2011-India
15. Shah N (2013) Harmonics in power systems causes, effects and control. Whitepaper design
engineering low-voltage drives
16. IEEE std 1531 (2003) IEEE guide for application and specifications of harmonic filter
Dual Fuzzy Direct Power Control
for Doubly-Fed Induction Generator:
Wind Power System Application
Abstract In this work, a control strategy based on fuzzy logic controllers is devel-
oped for the control of an asynchronous machine. The proposed control consists of
active and reactive powers estimation, two fuzzy controllers, and tracking based on
the theory of direct torque control. The originality of our work has been in the imple-
mentation of the different switching states in different fuzzy controllers by rules that
describe the operation of each magnitude controlled, ensured by the simplicity of
implementation and the robustness offered by the fuzzy logic.
As an approach, a fuzzy controller will be studied and implemented for the purpose
of improving the static and dynamic performance of a dual classical direct power
control applied to a wind power system. In this context, hysteresis controllers will
be replaced by two fuzzy controllers and the simulation results of the generator-side
and grid-side controls will be presented to prove the e of the control.
Nomenclature
Symbols Parameters
Vsα , Vsβ Stator voltages in (α, β) reference
Vgα , Vgβ Grid voltages in (α, β) reference
i sα , i sβ Stator currents in (α, β) reference
ir α , irβ Rotor currents in (α, β) reference
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1445
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_131
1446 H. Alami Aroussi et al.
1 Introduction
In many countries, wind power has become a pillar of their energy systems since it
is a reliable and inexpensive source of electricity.
Indeed, the advancement of using clean energies in recent years lead researchers
to develop wind energy controllers to exploit the maximum of wind energy extracted
and to ensure its quality.
Nowadays, the wind power system (WPS) based on a doubly-fed induction gener-
ator (DFIG) is the most implemented configuration. The latter has numerous advan-
tages that can be summarized mainly in the fact that the machine can drive under
variable speed, operates in both sub and hyper-synchronism modes, offers better
energy capture, reduces mechanical stress, and costs fewer thanks to the emerged
power converters [1–3].
The schematic of the wind power system configuration that we will use in the
simulation part is illustrated in Fig. 1. The DC-Link is linked to the back-to-back
converter. The rotor side converter (RSC) is connected to the rotor to control the
stator active and reactive powers while the grid side converter (GSC) is connected to
the grid to maintain the voltage of the DC-Link fixed [4].
As a first step, we propose a model of the wind power system, described before,
including the dynamic model of both the machine and the wind turbine. Then, we
choose the control that will be applied to run the whole system and will generate the
results displayed using the software Matlab/Simulink.
In our case, many techniques were developed by researchers, considering that
the system is linear or not, such as Field Oriented Control, Direct Power Control,
Backstepping, Sliding mode, Predictive Control…
Indeed, Vector Control is the most implemented one in wind farms. However, its
main drawback is the fact that this control is sensible to parameters variations of
the system such as resistance and inductance variations [5]. To overcome this, the
Direct Power Control, which is initially based on the Direct Torque Control, offers
direct control of the powers using a predefined lookup table based on estimation and
a perfect decoupling between reactive and active powers [6]. Though, the non-linear
Dual Fuzzy Direct Power Control for Doubly-Fed Induction Generator … 1447
behaviours of the WPS parts and the non-constant switching frequency lead us to
use non-linear techniques such as the Fussy Logic.
As a result, the major emphasis of the work is targeted at the application of Fuzzy
logic controllers as the replacement of the hysteresis controllers in the direct power
control scheme (Control level 1). The dual direct power control applied to the back-
to-back converter is carried under variable speed and the results are displayed through
simulations.
The dynamic model of the wind power system in the reference αβ can be expressed
by [7]:
The voltages of the doubly-fed induction machine, in space vector notation, are
written as:
1448 H. Alami Aroussi et al.
−
→ −
→ d ψ s Vsα = Rs i sα + dψsα
Vs = Rs i s + ⇒ dt
dψsβ (1)
dt
Vsβ = Rs i sβ + dt
−
→ −
→ d ψr −
→ Vr α = Rr ir α + dψr α
+ ωm ψrβ
Vr = Rr ir + − jωm ψr ⇒ dt
dψrβ (2)
dt
Vrβ = Rr irβ + dt
− ωm ψr α
The stator and grid powers (active and reactive) are given by the following equations
[4]:
−→ − →
Ps = 23 Re{Vs , i s∗ } ⇒ Ps = 23 (Vsα i sα + Vsβ i sβ ) (3)
−→ − →
Q s = 23 I m {Vs , i s∗ } ⇒ Q s = 23 (Vsβ i sα − Vsα i sβ ) (4)
−→ − →
Pg = 23 Re{Vg , i g∗ } ⇒ Pg = 23 (Vgα i gα + Vgβ i gβ ) (5)
−→ − →
Q g = 23 I m {Vg , i g∗ } ⇒ Q g = 23 (Vgβ i gα − Vgα i gβ ) (6)
1
Ptheor = .ρ.S.v 3 (7)
2
Though, the aerodynamic power appearing on the shaft of the turbine is expressed
as:
1
Paeor = C p .Ptheor = .ρ.S.C p (λ, β).v 3 (8)
2
The power coefficient C p (λ, β) will characterize the aerodynamic efficiency of
the turbine.
The previous equations will help us to elaborate a model of the main parts of our
wind power system. In the next task, we will explain the direct power control and
how to apply it to the machine part as well to the grid part.
Dual Fuzzy Direct Power Control for Doubly-Fed Induction Generator … 1449
The purpose of this technique (see Fig. 2) is to control directly the powers of our
system by applying different voltages to the converters. These vectors determine
their states [10, 11].
The two magnitudes controlled are the powers that are generally controlled by
hysteresis regulators.
The principle is to keep the magnitudes of the powers within these bands of
hysteresis. The performance of these comparators defines the optimum voltage to be
injected into the inverter for every switch [6].
The use of this type of regulator assumes the existence of a variable switching
frequency that needs a very minor calculation step. Furthermore, it is preferable
to operate with high frequency to decrease oscillations generated by this kind of
comparator.
Assuming that, in classic DPC, two types of power controllers are used: hysteresis
controllers that have a property based on the command and its history or strip
controllers where the command history does not play any role. They are based on
the number of power error bands [12].
In this approach, we will replace the classic regulators with two fuzzy controllers.
Here are all the rules describing the operating principle for this algorithm [13–15]:
⎧
⎪
⎪ If eQ s P then HQ I ncr ease
⎪
⎪
⎪
⎨ f
I eQ s N then HQ Decr ease
If e Ps P then H P I ncr ease (9)
⎪
⎪
⎪If
⎪ e Ps Z then H P is kept
⎪
⎩If e Ps N then H P Decr ease
In this part, the dual fuzzy direct power control applied to a wind power system of
1.5 kW was tested under the Matlab/Simulink environment with the presence of a
balanced three-phase load (Rload ) of 300 W.
The system, shown in Fig. 1, was driven under subsynchronous and hypersyn-
chronous speed (steady and transient state) regarding the following conditions:
• The generator operates at variable speed (see wind profile);
• The DC bus voltage: Vbus = 600 V.
The simulation results of the power fuzzy logic applied to dual DPC are given by
the following figures (Fig. 3):
Under the circumstances mentioned previously, we observe that the stator active
power follows its reference. It has a very fast dynamic and fewer harmonics.
Indeed, the figure above shows a very good response of the stator active and
reactive powers (around 0.28 s), where these follow their references perfectly with
almost zero static error except for the peak that appears at 3.9 s.
On the other hand, the grid active power presents a clear distortion beginning at t
= 3 s, while the grid reactive power transient is almost not seen due to a reasonably
decoupled control performance.
It should also be remembered that the reactive power setpoint is imposed zero to
obtain a unit power factor on both rotor and grid sides.
Also, the patterns of the stator, rotor, and grid currents show small harmonic rates.
These currents show also fewer ripples but keep their sinusoidal shapes.
Figure 4 shows the spectrums of the stator (2.05%), rotor (3.43%), and grid
(9.01%) currents. Notice that the Total Harmonic Distortion (THD) of the currents
is less than 10% recommended by the standards IEC 61727 and IEEE 1547.
Dual Fuzzy Direct Power Control for Doubly-Fed Induction Generator … 1451
10 600
8
Vbus (V)
V (m/s)
6 400
4
200
2
0 0
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
Time (s) Time (s)
(a) (b)
Stator Currents Zoom Stator Currents
2
2
1
0 0
-1
-2
-2
-4 -3
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 3.7 3.71 3.72 3.73 3.74 3.75 3.76 3.77 3.78 3.79 3.8
Time (s) Time (s)
(c)
Rotor Currents Zoom Rotor Currents
Zoom Rotor Currents (A)
6 6
Rotor Currents (A)
4 4
2 2
0 0
-2 -2
-4 -4
-6 -6
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 2.6 2.8 3 3.2 3.4 3.6 3.8 4 4.2
Time (s) Time (s)
(d)
Stator Active Power Stator Reactive Power
Stator Reactive Power (Var)
1000 2000
Stator Active Power (W)
Ps Qs
Psref Qsref
500 X: 0.2862
Y: -68.77 -1000 1000
0 -1200
X: 0.2998
Y: 0
-1400
-500 0
3.8 4 4.2
-1000
-1000
-1500
-2000 -2000
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
Time (s) Time (s)
(e) (f)
Fig. 3 Simulation results: (a) wind profile (b) DC voltage (c) abc stator currents (d) abc rotor
currents (e) stator active power (f) stator reactive power (g) abc grid currents (h) grid active power
(i) grid reactive power
1452 H. Alami Aroussi et al.
0 0
-5
-10 -5
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 3.7 3.71 3.72 3.73 3.74 3.75 3.76 3.77 3.78 3.79 3.8
Time (s) Time (s)
(g)
Grid Active Power Grid Reactive Power
1500
Pg Qg
Pgref Qgref
3000 1000
2000 500
1000
0
0
-500
-1000
-2000 -1000
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
Time (s) Time (s)
(h) (i)
Fig. 3 (continued)
a b c
Fundamental (50Hz) = 2.129 , THD= 2.05% Fundamental (5Hz) = 7.76 , THD= 3.43% Fundamental (50Hz) = 4.333 , THD= 9.01%
Isa Ira 6 Iga
0.4 5
5
4
0.3 4
3
0.2 3
2
2
0.1
1 1
0 0 0
0 200 400 600 800 1000 0 200 400 600 800 1000 0 200 400 600 800 1000
Frequency (Hz) Frequency (Hz) Frequency (Hz)
Fig. 4 THD: (a) stator current Isa (b) rotor current Ira (c) grid current Iga
5 Conclusion
The work performed is a digital simulation of the direct control of the powers of a
wind power system based on the doubly-fed induction generator. We have proposed
a method to improve the dynamic performance of direct power control by intro-
ducing a fuzzy type regulator while detailing its main modules such as Fuzzification,
Inference, and Deffuzification.
Dual Fuzzy Direct Power Control for Doubly-Fed Induction Generator … 1453
The simulation results on the machine and grid sides showed acceptable robustness
of these regulators, especially from the point of view of decoupling the active and
reactive powers, and the speed variation. The next work will be a comparison between
dual direct power control using a different type of comparators and its validation using
a testbench.
Appendix
References
1. Stiebler M (2008) Wind energy systems for electric power generation. Springer
2. Wu B, Lang Y, Zargari N, Kouro S (2011) Power conversion and control of wind energy
systems. Wiley, Hoboken
3. Derbel N, Zhu Q (eds) (2019) Modeling, identification and control methods in renewable energy
systems
4. Abad G, Lopez J, Rodríguez MÁ, Marroyo L, Iwanski G (2011) Doubly fed induction machine:
modeling and control for wind energy generation. Wiley-IEEE Press
5. Boubzizi S, Abid H, El Hajjaji A et al (2018) Comparative study of three types of controllers
for DFIG in wind energy conversion system. Prot Control Mod Power Syst 3:21
6. Alami Aroussi H, Ziani EM, Bouderbala M, Bossoufi B (2020) Improvement of the direct
torque control of the doubly fed induction motor under variable speed. Int J Power Electron
Drive Syst 11(1):97–106
7. Dehong X, Frede B, Wenjie C, Nan Z (2018) Advanced control of doubly fed induction
generator for wind power systems. The Institute of Electrical and Electronics Engineers. Wiley
8. Aissaoui AG, Tahour A (eds) (2016) Wind turbines - design, control and applications.
IntechOpen, no 4133
9. Bouderbala M, Bossoufi B, Alami Aroussi H, Taoussi M, Lagrioui A, Livinti P (2020) Dead-
beat control applied to wind power system. In: Proceedings the 5th international confer-
ence on renewable energy in developing countries, IEEE-REDEC’2020, March 24–26, 2020,
Marrakech, Morroco
1454 H. Alami Aroussi et al.
10. Alami Aroussi H, Ziani EM, Bossoufi B, Bouderbala M (2020) DPC & DNPC applied to wind
energy converter system. In: Proceedings of the 5th international conference on renewable
energy in developing countries, IEEE-REDEC’2020, March 24–26, 2020, Marrakech, Morroco
11. Lee Y-S, Martin HL (2018) Chow in power electronics handbook, 4th edn. Elsevier
12. Bossoufi B, Alami Aroussi H, Bouderbala M (2020) Direct power control of wind power
systems based on DFIG-generator (WECS). In: Proceedings the 12th international conference
on electronics, computers and artificial intelligence, IEEE-ECAI’2020, June 25–June 27, 2020
13. Sudheer H, Kodad SF, Sarvesh B (2018) Improvements in direct torque control of induction
motor for wide range of speed operation using fuzzy logic. J Electr Syst Inf Technol 5(3):813–
828
14. Chikouche TM, Hartani K, Bouzar S, Bouarfa B (2020) New direct power control based on
fuzzy logic for three-phase PWM rectifier. In: Hatti M (ed) Smart energy empowerment in
smart and resilient cities. ICAIRES 2019. Lecture Notes in Networks and Systems, vol 102.
Springer
15. Aissa O, Moulahoum S, Kabache N, Houassine H (2014) Fuzzy logic based direct power control
for PWM three-phase rectifier. In: 22nd mediterranean conference on control and automation,
Palermo, pp 79–84
DFIG-WECS Non Linear Power
Controls Application
Abstract This study proposes two different controlling approaches of Wind Energy
Conversion System (WECS). The first one deals with Field/Oriented/Control/ (FOC)
and the second with Direct Power Control (DPC). Both approaches were applied on
Doubly-Fed-Induction-Generator-(DFIG).
DPC seems to give good tracking and better efficiency compare to FOC. In fact,
by comparing Stator active and reactive powers, DPC gives much less oscillations.
Regarding rotor and stator currents, DPC needs less time to achieve steady-state.
1 Introduction
The field of electric machine control has known a great evolution in industrial sector.
Used for several centuries (windmills), wind energy has been developed considerably.
This significant development of this energy is going to be essential to attain the aims
set for renewable energies. A wind system is usually composed of turbine, gearbox,
and generator. There are two varieties of wind turbines: vertical axis wind turbines
and horizontal. The latter is the most used because of its great efficiency. Regarding
the generator, DFIG is often used thanks to its several advantages: the rotor power
reaches around 30% of the total power. As result, the dimensions of converters among
the conversion system is reduced [1].
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1455
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_132
1456 M. Bouderbala et al.
For many years, power control has been the topic of many researches. Many
control strategies have been proposed. Among these strategies, “vector controls”
gives dynamic performance identical to that acquired by the-direct-current (DC)
machine. Nowadays, the improvement of modern digital signal processing strategies
has prompted extensively, to give a rise to “Direct Control of the Powers” [2].
Blaschke was the first one to develop the primary theory of the field-oriented
techniques in 1970. Ten years later, its effective applications were adopted. Indeed,
it requires calculations of Park transform, evaluation of trigonometric functions and
regulations, which could not be done in an analogical approach. Vector control
is designed to provide the same control performance as a separately excited DC
machine, which is characterized by a natural decoupling of the variable flux and
torque [3].
In the middle of the 1980s, TAKASHI initiated the Direct control methods of
asynchronous machines as concurrent with classical controls depending on Pulse-
Width-Modulation (PWM) which uses field’s orientation to guarantee the decoupling
between flux and torque. This study deals with direct power control (DPC) that lies
on the principle of direct torque control (DTC) [4].
These two control strategies discussed previously both aims to control
the/active/and/reactive/powers/separately.
The comparison between FOC and DPC evaluate the difference between two
different regulators: PI and hysteresis. However, each control lies on a different
approach. Vector control is based on PI controllers, which are highly dependent
on the machine parameters, while DPC control is based on hysteresis/comparators,
which do not depend on the machine parameters.
In order to compare between DPC and FOC control strategies, this work is orga-
nized as follows: First, In Sect. 2 the turbine and DFIG was modelled to model
the WECS. The Field Oriented Control and Direct Power Control are described
in Sect. 3 and Sect. 4 respectively. The results simulations are obtained using
Matlab/Simulink then, an extensive results discussions are presented in (Sect. 5).
Finally, the conclusion is presented.
Figure 1 illustrates the synopsis of the studied system. The/wind/turbine drives the
DFIG at a variable rotational speed through a multiplier. The DFIG stator is linked
to the network directly. On the other hand, the rotor is linked to the network via
a/back-to-back/converter and transformer [1].
First of all, it is necessary to model the system before proceeding to control.
DFIG-WECS Non Linear Power Controls Application 1457
Stator Power
Grid 50Hz
Wind
DFIG
turbine
Rotor
Controls
The aerodynamic/power (Paer ) is the product of power coefficient (Cp ) and wind
power (Pv ). This power is calculated from wind speed (V), blades swept area (S),
and air density (ρ).
The power coefficient presents efficiency as a function of beta angle (β) and the
lambda ratio (λ). The ratio depends on the wind speed (V), blade-radius (R), and
turbine speed () t [2].
The following equations describe wind turbine system.
ρ.S.V3
Pv = (2)
2
R.t
λ= (3)
V
The multiplier is described by the equations cited below:
Ct
Cg = (4)
G
mec
t = (5)
G
With, Cg, mec and G are generator/torque, mechanical/speed of the generator,
and multiplier/gain respectively.
The dynamic equation of the turbine is given by:
1458 M. Bouderbala et al.
dmec
J = Cmec = C g − Cem − C f .mec (6)
dt
With, Cem, Cmec, Cf, and J: electromagnetic torque, mechanical torque, and
viscous friction torque, total inertial respectively.
There are three types of equations organize the dynamic modeling of DFIG in the
park frame (d.q): Electrical equations, Flux equations and electromagnetic torque
[5].
Electrical equations:
dϕsd
V sd = Rs.I sd + − ϕsq.ωs (7)
dt
dϕsq
V sq = Rs.I sq + + ϕsd.ωs
dt
dϕr d
V r d = Rr.I r d + − ϕrq.ωr
dt
dϕrq
V rq = Rr.I rq + + ϕr d.ωr
dt
Flux equations:
ϕr d = Lr.I r d + Lm.I sd
Electromagnetic torque:
M
Cem = p. . ϕsq .Ir d − ϕsd .Irq (9)
Ls
Vs (d.q), ϕs (d.q), Is (d.q): voltages, flux, and currents statoric in PARK reference
Vr (d.q), ϕr (d.q), Ir (d.q): voltages, flux, and currents rotoric in PARK reference
Rs, Rr: stator and rotor resistances
DFIG-WECS Non Linear Power Controls Application 1459
ωr = ωs − ω (10)
ω = p. (11)
As result, the DFIG can easly be controlled. Active power is controlled separately
from reactive power [6, 7].
It is assumed that the stator resistance is negligible as well as the flux is constant.
Equations 7 becomes then:
Vsd = 0 (13)
Vsq = ϕsd.ωs
Injecting Eqs. 12 and 13, in Eqs. 8 and we obtain the following equations of stator
currents and stator flux respectively:
ϕs M
Isd = . .Ir d (14)
Ls Ls
M
Isq = − .Irq
Ls
1460 M. Bouderbala et al.
The following equations are obtained using stator currents (Eqs. 14) and stator
voltages (Eqs. 13).
3 M
Ps = − . .V s.I rq (16)
2 ϕsd
3 ϕsd M
Qs = .V s − .I r d
2 Ls Ls
M 2 d Ir d M2
Vr d = Rr .Ir d + (L r − ). − g.ωs .(L r − ).Iqr
Ls dt Ls
(17)
M 2 d Irq M2 M.Vs
Vrq = Rr .Irq + (L r − ). − g.ωs .(L r − ).Idr + g.
Ls dt Ls Ls
The switching table presented in this section divides the plan into six sectors ((1)–(6)).
Other tables divide the plan into twelve sectors to improve accuracy.
When the voltage vector is aligned with the chosen rotating reference frame,
the power is related to the direct component of the output current and the power
instantaneous reactivation to the quadrature current [13].
The selection of the inverter switching mode is imposed by two hysteresis bands
Hp, Hq in order to maintain the errors between the power reference values (Pref and
Qref) and their measured values in these bands.
To accomplish this, the errors of the instantaneous active and reactive powers are
processed by the hysteresis comparators. Their outputs are set to 1 to increase the
control variable (P or Q) and to 0 when the variables controlled are unchanged or
must decrease [14, 15].
The hysteresis controllers are used for the purpose to adjust the active and reactive
power settings.
One is influenced by the error ep = Pref − P for active power. The other is
influenced by the error eq = Qref − Q for reactive power.
The hysteresis band widths affect the efficiency of the inverter, especially, the
distortion of the harmonic current, and the frequency average switching time [16, 17].
The following Synopsis Fig. 3 presents the DPC modelling through the hysteresis
comparators then, vector selection after that the pulse generation as indicated above:
1462 M. Bouderbala et al.
5 Results
To approve these controls, Matlab/Simulink was used. The results identified with a
DFIG of 2MW are given as follows.
10
6 / / 10
6
Stator Reactive Power ( Var)
2 2
Stator Active Power ( w)
1
1
0
0
-1
-1
-2
-2
1 2 3 4 5
1 2 3 4 5
times(s)
times(s)
5000 2000
Rotor Currents (A)
0
0
-2000
-5000
-4000
0 1 2 3 4 5
1 2 3 4 5 times(s)
times(s) 4000
Stator Currents (A)
-2000
-4000
4.3 4.4 4.5 4.6
times(s)
d Stator Currents
6 6
10 10
2 2
1 1
0 0
-1 -1
-2 -2
1 2 3 4 5 1 2 3 4 5
times(s) times(s)
5000
2000
0 0
-2000
-5000
-4000
1 2 3 4 5 1 2 3 4 5
times(s) times(s)
c Rotor Currents
Stator Currents(A)
2000
-2000
d Stator Currents
Both FOC and DPC control strategies are simulated, tested, and compared to their
powers reference. Due to this reason, steps have been applied as references.
The active power step varies from 0 to −2 MW at times t = 4 s; while the reactive
power step varies also from 0 to −2 MVAR at times t = 25 s. The simulation results
are presented in Fig. 4 for the FOC strategy and Fig. 5 for the DPC strategy; as
follows (a): active power Ps, (b): reactive power Qs, (c): rotor current Ir, (d): stator
current Is.
Firstly, based on Fig. 4a, b and Fig. 5a, b we can clearly conclude that; active and
reactive powers are decoupled. Also, the variables follow their references for both
types of controls. At the instant t = 2 and t = 4, the currents Fig. 4c, d and Fig. 5c,
d change values with the variations of powers Fig. 4a, b and Fig. 5a, b respectively.
By comparing Stator active Fig. 4a and Fig. 5a, and reactive powers Fig. 4b and
Fig. 5b, DPC gives much less oscillations. Regarding rotor and stator currents, DPC
needs less time to achieve steady-state.
From the results, it can be deduced that with hysteresis comparators, there are
less oscillations even the response time is reduced compared to the PI controllers.
1464 M. Bouderbala et al.
The Table 1 summarizes the principal differences between the FOC and the DPC
strategy.
6 Conclusion
This work was dedicated to a comparative study between field-oriented control (FOC)
and direct power control (DPC) applied to a wind energy conversion system (WECS).
The configuration and principle of the FOC and the DPC has been detailed.
The FOC is based on the PI regulator. This regulator has some drawbacks among
them it depends greatly on machine parameters. Though, The DPC is based on
hysteresis comparators and independent of parameter variations.
In this way, the authors can attest that the utilization of the hysteresis comparators
in DPC control has different focal points, for example, diminishing the exchanging
recurrence of intensity switches and improving the waveforms of the yield factors of
the machine, instead of the utilization of the corresponding fundamental regulator
(PI) in the FOC.
Thus, the authors can affirm that the use of the hysteresis comparators in DPC
control has various advantages such as: decreasing the switching frequency of power
switches and improving the waveforms of the output variables of the machine, rather
than the use of the proportional-integral controller (PI) in the FOC.
The results achieved prove the effectiveness of the DPC strategy employed for
DFIG-based system control and attest to the expected performance.
After obtaining these results in the Matlab software, as a perspective, we envisage
applying these commands on the test bench.
Funding The authors declare that they have funding for research from CNRST.
Conflict of Interest The authors declare that they have no conflict of interest.
DFIG-WECS Non Linear Power Controls Application 1465
References
1. Gonzalo A, Jesus L, Miguel R, Luis M, Grzegorz I (2011) Doubly fed induction machine:
modeling and control for wind energy generation. Wiley, Hoboken
2. Bossoufi B, Aroussi H, Bouderbala M (2020) Direct power control of wind power systems
based on DFIG-generator (WECS). In: Proceedings of the 12th international conference on
electronics, computers and artificial intelligence, IEEE-ECAI’2020, June 25–June 27, 2020
3. Li S, Wang H, Tian Y, Aitouch A, Klein J (2016) Direct power control of DFIG wind turbine
systems based on an intelligent proportional-integral sliding mode control. ISA Trans 64:431–
439
4. Alami Aroussi H, Ziani EM, Bouderbala M, Bossoufi B (2020) DPC & DNPC applied to wind
energy converter system. In: Proceedings of the 5th IEEE-REDEC’2020, March 24–26, 2020,
Marrakech, Morroco
5. Bouderbala M, Bossoufi B, Aroussi H, Lagrioui A, Taoussi M, Livinti P (2020) DEAD-
BEAT control applied to wind power system. In: Proceedings of the 5th international confer-
ence on renewable energy in developing countries, IEEE-REDEC’2020, March 24–26, 2020,
Marrakech, Morroco
6. Siegfried H (2006) Grid integration of wind energy conversion systems. Wiley, Chichester
7. Muller S, Deicke M, De Doncker RW (2003) Doudly fed induction genertor systems for wind
turbines. IEEE Ind Appl Mag 8:26–33
8. Alami AH, Ziani EM, Bouderbala M, Bossoufi B (2020) Improvement of the direct torque
control of the doubly fed induction motor under variable speed. Int J Power Electron Drive
Syst 11(1):97–106
9. Noguchi T, Tomiki H, Kondo S, Takahashi I (1998) Direct power control of PWM converter
without power-source voltage sensors. IEEE Trans Ind Appl 34(3):473–479. https://doi.org/
10.1109/28.673716
10. Kendouli F, Boulahia A, Nabti K, Benalla H (2010) Modeling of a variable speed wind turbine
driving PMSG using hysteresis band pulse width modulation (PWM). Int Rev Model Simul
3(6):1196–1201
11. Kumar DK, Ram Das GT (2018) Simulation and analysis of modified DTC of PMSM. Int J
Electr Comput Eng 8(5):2894–2903
12. Alami AH, Ziani EM, Bouderbala M, Bossoufi B (2020) Enhancement of the direct power
control applied to DFIG-WECS. Int J Electr Comput Eng (IJECE) 10(1):35–46
13. Tamalouzt S, Idjdarene K, Rekioua T, Abdessemed R (2016) Direct torque control of wind
turbine driven doubly fed induction generator. Rev. Roum. Sci. Techn.– Électrotechn. et Énerg.
61(3):244–249
14. Daoud A, Derbel N (2019) Direct power control of DFIG using sliding mode control approach.
In: Derbel N, Zhu Q (eds) Modeling, identification and control methods in renewable energy
systems. Green Energy and Technology. Springer, Singapore
15. Kairous D, Belmadani B (2015) Robust fuzzy-second order sliding mode based direct power
control for voltage source converter. Int J Adv Comput Sci Appl 6:167–175
16. Yousefi-Talouki A, Zalzar S, Pouresmaeil E (2019) Direct power control of matrix converter-fed
DFIG with fixed switching frequency. Sustainability 11:2604
17. Monmasson E, Cirstea M (2007) FPGA design methodology for industrial control systems – a
review. IEEE Trans Ind. Electron. 54(4):1824–1842
Behaviour of Concrete and Steel
for the Construction of Storage Tanks:
Comparative Study
Abstract The demographic growth and economic development are two main factors
which create more demand for energy reserve production in petrochemical industry.
Under such circumstances, an increase in the volume and quantity of storage tanks has
become must due to the their importance in preventing harms and risks by storing
large volumes of flammable, explosive, toxic and harmful materials, that is why
many researchers are orientating themselves in research projects of materials for the
construction of storage tanks with good quality and with the cheapest. In this article,
we show the different thermo-mechanical behaviors of the two materials which are
concrete and steel so that the constructors can make the most adequate choice for
their needs (storage conditions, temperature, pressure, etc…), especially since each
of these materials has specific characteristics that can influence the types of tanks as
well as the storage modes.
1 Introduction
In recent years, several events have led to the development of the construction of
storage tanks, to store the various products encountered in the petrochemical and
chemical industry, which is mainly carried out in metal tanks, which are welded
and installed freely in the air (overhead tanks) [1]. Storage has been the subject of
much attention because it offers the possibility of energy savings and efficient use
[2]. For this reason, manufacturers use various types of storage tanks, for example,
aboveground tanks, flat-bottomed tanks, concrete silos or steel for the storage of coal,
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1467
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_133
1468 S. El Aamery et al.
coke, grain, etc., tanks spherical (pressure vessels) for the storage of water and oil, of
high-pressure liquefied water and underground tanks for the storage. This requires
thinking about tanks with larger volumes [3], therefore, it is useful to do research,
with the aim of comparing the properties, and the thermo-mechanical behaviour of
the materials, in order to select the best to build the storage tanks, the most efficient
at the lowest cost [4].
Numerous researchers are studying the physical and chemical properties of
concrete, as concrete is a material widely used all over the world and in different
fields [5]. It is composed of four basic ingredients: water, cement, gravel as coarse
aggregate, and sand as fine aggregate. On the other hand, advanced energy storage
systems are being actively pursued worldwide to respond to the rapid development of
sustainable energy [6, 7]. Aggregates are the world’s most mined mineral resources,
and the global construction market consumed the equivalent of $360 billion in 2018
alone [7]. In this context, this article also discusses concrete as a building material
for storage tanks where it will be compared with steel.
For a better presentation of the results and the different characteristics of these
materials, our article is organized as follows: A review of the literature on tanks
and storage methods is described in Sect. 2. Section 3 unveils a comparative study
between concrete and steel. A discussion is presented in Sect. 4; finally, a conclusion
will be presented in the last section.
2 Literature Review
There are several structures in the industrial sector, among the best known of these
structure, there are storage tanks, which are of different types: vertical cylindrical
tanks, horizontal cylindrical tanks, spherical tanks and spheroidal tanks, each of these
tanks has its own specific characteristics as they can be used for different purposes
in industrial projects [8].
Storage tanks Liquid, as special structures, behave differently from ordinary struc-
tures in terms of dynamic characteristics [9]. They are most often used in distribution
systems, industries and nuclear power plants for the storage of drinking water, toxic
and flammable liquids and nuclear waste respectively. Because earthquake damage
to liquid storage tanks can cause secondary hazards, they are strategically important
structures [10], ich are built in two different shapes, either cylindrical or rectangular.
Figure 1 shows the cylindrical tank with the chordal system (r; 8; z).
Behaviour of Concrete and Steel for the Construction of Storage Tanks … 1469
The dimensions of the tank and its material depend on the fluids to be stored and
the type of storage. Oil storage generally consists of a steel structure for a shell and
annular reinforced concrete beams that transfer the load to the ground. The storage
mode can be classified according to the pressure and the operating temperature of
the stored liquid, taking into account the relationship between these two parameters.
Figure 2 shows the three storage modes for liquids with a normal boiling point
(STP) below room temperature.
Fig. 2 The three storage modes for liquids with a normal boiling point (STP) below room
temperature [1]
1470 S. El Aamery et al.
cement
[POURCENT
AGE]
water
[POURCENT
AGE]
Steel is a material consisting mainly of iron and some carbon, which are extracted
from natural raw materials from underground (iron and coalmines). Carbon is only
involved in the composition of the material for a very small part (generally less than
1%) [15]. In Table 1, we will show the symbols and values of the steel properties.
On the other hand, electrical steels used for motors, transformers and generators
are usually covered with an insulating coating to improve the performance of the
material in terms of reduced power loss, punching and welding characteristics and
corrosion resistance [16].
Behaviour of Concrete and Steel for the Construction of Storage Tanks … 1471
In general, the tensile strength and the flexural strength of concrete are respectively
of the order of ~10 and ~15% of the compressive strength. In addition, compressive
strength is often considered as a main quality of concrete because it is directly related
to the structure of the hydrated cement paste [17]. On the other hand, concrete is
classified in several strength classes based on compressive strength in most European
standards [18].
As an example, fiber reinforced concrete (FRC) is a high performance material
for the construction and reinforcement of structural elements exposed to impacts
and other types of extreme loads. Many studies for prediction were a reference to
the advantages of incorporating fibers in concretes exposed to explosions [19, 20],
ballistic shocks [21], cyclic loads, and different types of impacts [22]. When fibers
are homogeneously distributed and used in appropriate quantities inside the concrete,
they reduce strength, toughness, ductility, durability and improve other mechanical
properties [23].
Several authors have studied the dynamic behavior of steel storage tanks and their
contents using simplified models capable of predicting the seismic response of these
facilities [24]. These tanks are generally used in petrochemical plants.
Vertical cylindrical steel storage tanks are responsible facilities that are filled
to their maximum capacity for most of their operational life. They often contain
toxic, flammable, and/or expensive products and must therefore be treated with care
and responsibility [25]. They can also be stored above ground using sandbags or
tires as chocks [26]. Steel structures must have sufficient toughness to withstand
various operating temperatures [27], that is why several tests were done to reveal its
1472 S. El Aamery et al.
Fig. 4 Diagrammed
contrainte/deformation
mechanical characteristics, the most important one is the tensile test that leads to the
following diagram (stress/strain) [28] (see Fig. 4).
• OA: rectilinear zone where there is proportionality between the deformation and
the applied stress. It is a reversible elastic zone
• AA’: horizontal bearing, reflecting elongation under constant load. There is
material flow, this is the plastic zone.
• A’B: the load increases again with the elongations up to point B. If the specimen
is unloaded in the plastic zone A’B a remanent elongation is observed; if it is
reloaded, an elastic behavior up to the previous load is observed: the yield strength
has been increased, the metal has been hardened.
• BC: Elongation continues, although the load is decreasing, until point C where it
breaks. In this phase, there is striction: the plastic deformation is localized in a
small portion of the specimen and is therefore no longer homogeneous.
When tensile tests were done on a steel beam and a concrete beam, with two different
temperature conditions, to compare the displacements of these two materials, we find
that:
At a temperature of 100 °C: we notice that the displacements of these two materials
increase (see Fig. 5), also the displacements of steel are a little bit larger than concrete
of the order of 0.02, same thing we notice at a temperature of 500 °C (see Fig. 6).
Thus, the displacements of either concrete or steel are higher at 500 °C than at
100 °C. Thus, we can conclude that the displacement increases with the increase in
temperature.
Behaviour of Concrete and Steel for the Construction of Storage Tanks … 1473
0.6
0.5
0.4
0.3
0.2
0.1
0
0 100 200 300 400 500 600
steel concrete
2.5
1.5
0.5
0
0 100 200 300 400 500 600
steel concrete
4 Discussions
In the world of energy storage, we are always looking for the evolution and use of
materials and storage conditions, in this context that always researchers try and carry
out studies for new materials and compare them in order to make the best choice in
terms of price quality.
In this part, we will compare two materials that are Steel S355 and Concrete B30, to
know their thermo-mechanical behavior and study the advantages and disadvantages,
to build storage tanks. Table 2 and Table 3 below show the different properties of
strengths and stiffnesses in MP.
1474 S. El Aamery et al.
Compared to concrete tanks, steel tanks have many advantages which are:
Main Advantages
• The high tensile strength of steel.
• The good resistance to earthquakes.
• Easy transportation due to the low weight which allows to transport far.
• Architectural possibilities, much more extensive than concrete [29].
On the other hand, this material also has a number of disadvantages, of which one
finds.
Main Drawbacks:
• Lower compressive strength than concrete.
• Poor fire resistance, requiring expensive protective measures.
• Need for regular maintenance against corrosion.
5 Conclusion
Based on this comparative study between concrete and steel, we can conclude that
concrete is a good material, which resists compression unlike steel, which is more
rigid in traction. Therefore, to better build a storage tank, we recommended using a
composite material of concrete and steel, or using a high performance concrete, espe-
cially if it is used with high temperatures and high pressures, that is to say, physical
properties that are particularly suitable for municipal and industrial applications.
As a research perspective, we are thinking of modeling by the finite element
method the storage tanks built by any material in order to have the thermo-mechanical
behaviors.
Behaviour of Concrete and Steel for the Construction of Storage Tanks … 1475
References
21. Almusallam TH, Siddiqui NA, Iqbal RA, Abbas H (2013) Response of hybrid-fiber reinforced
concrete slabs to hard projectile impact. Int J Impact Eng 58:17–30. https://doi.org/10.1016/j.
ijimpeng.2013.02.005
22. Yahaghi J, Muda ZC, Beddu SB (2016) Impact resistance of oil palm shells concrete reinforced
with polypropylene fibre. Constr Build Mater 123:394–403. https://doi.org/10.1016/j.conbui
ldmat.2016.07.026
23. Mehta PK, Monteiro PJM (2001) Concrete microstructure, proprietis and materials, 20 Oct
2001
24. Caprinozzi S, Paolacci F, Dolšek M (2020) Seismic risk assessment of liquid overtopping in a
steel storage tank equipped with a single deck floating roof. J Loss Prev Process Ind 67:104269.
https://doi.org/10.1016/j.jlp.2020.104269
25. Elsevier Enhanced Reader. https://reader.elsevier.com/reader/sd/pii/S2452321620300913?
token=BAA4AC4C9E8559789F2173771B34EB6C910AFAAD131CA7A1AB38CBBBD
7B9982BA183C696752D3A40375F3FBFE9D06392. Accessed 25 Sep 2020
26. Cheremisinoff NP, Cheremisinoff PN (1995) Steel and fiberglass construction for below ground
storage tanks. In: Fiberglass reinforced plastics. Elsevier, pp 158–180
27. Diagnosis of the microstructural and mechanical properties of over century-old steel railway
bridge components. Elsevier Enhanced Reader (2020). https://reader.elsevier.com/reader/
sd/pii/S1350630719311501?token=406588265883F448006A629FB71499E9200E0552B3
C37465EB2E1FAE3EBF2B8B895A70028F5D6F5C811518653F7C41AF. Accessed 08 Oct
2020
28. Les essais sur le matériau acier. https://notech.franceserv.com/materiau-acier-essais.html.
Accessed 09 Oct 2020
29. Tomek R (2017) Advantages of precast concrete in highway infrastructure construction.
Procedia Eng 196:176–180. https://doi.org/10.1016/j.proeng.2017.07.188
Sensorless Control of DC-DC Converter
Using Integral State Feedback Controller
and Luenberger Observer
Abstract This paper presents a design of linear state feedback control of DC-DC
Boost converters, in order to achieve a particular desired behavior. To guarantee a
zero steady state error, we introduce an integral action, which will work out this
problem by assuring that the steady state error will end up to zero. If it is supposed
that both the voltage and current are measured, so much more sensors are needed
then and consequently causing a high cost, so that to estimate the voltage and current
with a low cost and less complexity it is preferred to introduce a state observer. An
observer or estimator is a dynamic system that uses the available information on a
real system, according to the inputs and outputs of the real process and estimate the
system state. Simulations results demonstrate the robustness and effectiveness of the
proposed control scheme.
D. Taibi (B)
Département de Génie Électrique, Université de Kasdi Merbah, Ouargla, Algeria
T. Amieur
Département de Génie Électrique, Université de Larbi Tebessi, Tebessa, Algeria
e-mail: [email protected]
T. Amieur · M. Bechouat · M. Sedraoui
Laboratoires des Télécommunications LT, Université 8 mai 1945, Guelma, Algeria
e-mail: [email protected]
M. Bechouat
Faculté des Sciences et Technologie, Université de Ghardaïa, Noumirat BP 455,
Route Ouargla Ghardaia, 47000 Ghardaia, Algeria
S. Kahla
Centre de Recherche en Technologies Industrielles CRTI, 16014 Cheraga Alger, Algeria
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1477
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_134
1478 D. Taibi et al.
1 Introduction
The boost converter has earned its name due to its ability of producing a DC output
voltage greater in magnitude than the DC input voltage (step-up). The Circuit diagram
of a Boost converter is as shown in Fig. 1.
While the transistor is passing current, the diode is reverse biased. Hence, the
source E is not connected to a load R.
While the transistor is blocking current, the diode is forward biased and
conductive, which makes the source E connected to a load R [12–14].
Sensorless Control of DC-DC Converter … 1479
From the two Eqs. (1) and (2), we can obtain a single unified model, which is
u ∈ {0, 1}. So the dynamic of the converter is as follows [12]:
L di
dt
= E − v(1 − u)
(3)
C dt = − Rv + i(1 − u)
dv
2.2 Linearization
The State feedback controller consists in considering the model of the process
described by the following equation [15]:
ẋ(t) = Ax(t) + Bu(t)
(5)
y(t) = C x(t)
It assumes xi (t) of the state vector x(t) are accessible for measurement.
A possible state-feedback control law can be given as [16]:
ẋ(t) = (A − B K c )x(t) + Br (t)
(7)
y(t) = C x(t)
As we have seen before, the status feedback command moderates the poles of the
system in closed loop. However, the latter does not ensure a position error. One
possibility to solve the problem is to add an integrator in the direct chain as shown
in Fig. 5.
We introduce an integral action: a noted state variable xi placed at the exit of the
integrator with [18].
xi = (r (t) − y(t)) = (r (t) − C x(t)) (8)
1482 D. Taibi et al.
Where v(t) is an external disturbance which acts on the states of the system
through the matrix B0 . The command designed before the screen cannot cancel the
effect of this disturbance on monitoring the reference which is given by:
⎧
⎨ ẋ(t) = Ax(t) + Bu(t) + B0 v(t)
ẋi (t) = −C x(t) + r (9)
⎩
y(t) = C x(t)
If we define the augmented state vector xaT = x xi , we can write the system in
the following increased form [18]:
ẋa (t) = Aa xa (t) + Ba u(t) + B0a v(t)
(10)
y(t) = Ca xa (t)
Aa = , Ba = , B0a = , Ca = C 0
−C 0 0 0
With
A0
Aa = (17)
C0
4 Luenberger Observer
In this part, it is considered that all the state variables are available for the feedback.
However, practically, some variables may not be available. Therefore, the estimate
phase of those variables is definitely an essential step, means, that we must estimate
the non-measurable state variables so as to elaborate the control signals.
The observer is given by [19, 20]:
ˆ = A x̂(t) + Bu(t) + L y − ŷ
ẋ(t)
(18)
ŷ(t) = C x̂(t)
L: gain, L T = l1 l2 . . . ln .
Error is defined as
˙
ė(t) = ẋ(t) − x̂(t) (20)
ė(t) = A x(t) − x̂(t) + L y(t) − ŷ(t) (22)
The converter is controlled by the PWM technique (carrier and modulating) linked
to an open loop IGBT switch (Figs. 7, 8 and 9).
Discussion
– The controlled system in open loop is a blind system.
– The open loop simulation gives a short rise time, but a long response time with
oscillations and in addition a significant error and this is not good for the system
(Figs. 10, 11, 12 and 13).
Discussion
– State feedback improves system stability and eliminates overflow and static error
of the system that appears in the open loop.
– The integral action compensates the error that the variation of the load creates.
0.8
0.6
0.4
0.2
0
0 0.005 0.01 0.015 0.02 0.025 0.03 0.035 0.04 0.045 0.05
Time (sec.)
20
15
10
-5
0 0.005 0.01 0.015 0.02 0.025 0.03 0.035 0.04 0.045 0.05
Time (sec.)
– According to the gotten results from the above graphs, the convergence of the
estimated current and voltage towards its true value is perfectly achieved.
1486 D. Taibi et al.
0.8
0.7
Control signals
0.6
0.5
0.4
0.3
0.2
0.1
0 0.02 0.04 0.06 0.0 8 0.1 0.12 0.14 0.16 0.18 0.2
Time (sec.)
Fig. 12 Current 1
I Estimate
0.9 I Measurement
0.8
0.7
0.6
Current
0.5
0.4
0.3
0.2
0.1
0
0 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 0.2
Time (sec.)
Sensorless Control of DC-DC Converter … 1487
Fig. 13 Voltage 25
20
15
Voltage
10
5
V Estimate
V Measurement
Reference
0
-5
0 0.02 0.04 0.06 0.08 0.1 0.12 0.14 0.16 0.18 0.2
Time (sec.)
6 Conclusion
This paper has demonstrated the control design methods for Boost converter. The
state feedback control for with an integrator, underling the clear advantages of state
feedback, that has a positive effect on response settling time, reducing the undesirable
peak overshoots and serve having a less oscillated performance, referring that this
approach doesn’t provide a zero static error, this latter has been solved by adding an
integral action to state feedback control, that has proven its efficiency when working
out the steady-state error then the use of state estimation technique with Luenberger
observer was discussed and simulated, that comes out with the advantages of the
observer in estimating those non measurable state variables.
References
1. Hsu CF, Lee TT, Wen YW, Ding FS (2006) Intelligent control for DC-DC power converter
with recurrent fuzzy neural network approach. In: 2006 IEEE international conference on fuzzy
systems, pp 457–462
2. Xia C, Geng Q, Gu X, Shi T, Song Z (2012) Input–output feedback linearization and speed
control of a surface permanent-magnet synchronous wind generator with the boost-chopper
converter. IEEE Trans Ind Electron 59(9):3489–3500
3. Ortiz-Lopez MG, Leyva-Ramos J, Carbajal-Gutierrez EE, Morales-Saldana JA (2008)
Modelling and analysis of switch-mode cascade converters with a single active switch. IET
Power Electron 1(4):478–487
4. Woo YJ, Le HP, Cho GH, Cho GH, Kim SI (2008) Load-independent control of switching DC-
DC converters with freewheeling current feedback. IEEE J Solid-State Circuits 43(12):2798–
2808
5. Jaen C, Pou J, Pindado R, Sala V, Zaragoza J (November 2006) A linear-quadratic regulator
with integral action applied to pwm dc-dc converters. In: IECON 2006–32nd annual conference
on IEEE industrial electronics. IEEE, pp 2280–2285
1488 D. Taibi et al.
Abstract Nowadays, modern drying methods have replaced traditional drying that
consists of exposing the product to direct sunlight. The indirect solar dryer is one
of the best-known systems used since this revolt. For the latter, there are different
concepts and equipment to build it, and although there are differences between them,
they all have the same goal, which is preserving food by reducing its moisture content.
These methods use heat under appropriate conditions in terms of temperature, relative
humidity, and air velocity, which results in a decrease in the moisture content of the
dried product, in the extent to prevent enzymatic activities as well as all biological
organisms that can be the cause of a corruption of the vegetal products. Therefore, it
is very important to ensure a good dried product, for that several parameter must be
adjusted, the objective is to optimize the distribution of the airflow along the trays
and to improve the efficiency of the system based on the dimensional parameters and
the quantity of the product that can be dried, for that, the optimal number of trays
and the distance between the trays is investigated using ANSYS FLUENT software
under the oriental region climate of Morocco. The results show that a drying chamber
with 3 trays found to be the optimal number with a distance of 5 cm between the
trays.
1 Introduction
There have been many ways of preserving organic products in different parts of the
world thousands of years ago, and these methods differ depending on the product to be
dried and the regional traditions. Some of the most knowns methods of preservation
are drying, canning, freezing, processing (smoking or salting), and pickling.
Morocco has considerable solar potential that present for most of the year, which
makes the solar drying the most widely used method for preserving products. But,
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1489
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_135
1490 D. Chaatouf et al.
after the failure of traditional drying [1] that consists of drying the products in the open
air, extensive research has been carried out to design and develop the performance
of solar dryers, some of them are presented in this section.
The general design of a drying system is mainly based on the dimensions and the
construction concept of the most important components such as the drying chamber,
the solar air heater, and other several parameters that influence the drying quality
of the system. However, the chamber geometry and the solar air heater have not
interested some developers who have chosen to study how to encourage airflow
through intake fans driven by photo-voltaic (PV) solar panels [2, 3]. In addition,
several studies were made on the hybridization of the indirect solar dryer, like the
one made by S. Dhanushkodi et al. [4] in which they use biomass and solar energy as
sources. The study of the dryer was also investigated on different modes, to idealize
and analyze the behavior of the system. There are many modes of operation to
evaluate the efficiency of the dryer, I. Montero et al. [5] conducts a comparison study
between indirect, passive, mixed, active, forced convection, natural convection and
hybrid modes of operation, and it has been revealed that the forced hybrid is the most
efficient mode of operation, followed by passive and active modes.
In the indirect natural convection solar dryer, the most common problem with it
is the non-uniform distribution of air inside the chamber, which affects the quality
of the dried products. Therefore, in this study, we will focus on the geometric study
and optimization of the dimensional parameters specifically, the distance between
the trays and the number of it which is a crucial parameter in terms of the uniformity
of air [6]. The chosen geometry in this paper is an electric solar dryer studied by
Y. Amanlou et al. [7] that characterize with a side mounted plenum chamber, which
affects positively on the air distribution inside the chamber, but in our case which is
natural convection solar dryer equipped with a solar air heater, the distribution of air
is going to be totally different, that we need a new study to ensure good uniformity
of air along the trays.
2 Methodology
In this study, the drying system presents an indirect natural convection solar dryer
(Fig. 1) consists with a drying chamber where the figs are distributed on trays as
a product to be dried, equipped with a trapezoidal plenum chamber and a solar air
heater of a 1.5 m2 that composed of an aluminium absorber and simple glazing.
Optimization and Improvement of the Efficiency of a Drying System … 1491
Solar radiation
Air inlet
Fig. 2 Meteorological
station used to collect the
data
The meteorological data used in this paper are collected from the high precision
station shown in Fig. 2 installed in the university Mohamed first Oujda. The main
data provided for this study are the ambient temperature and the solar irradiation
presented in Eqs. (1) and (2) respectively, that have been used in the subroutine UDF
(User Defined Function) that we wrote to integrate it in ANSYS FLUENT so we can
study the solar dryer more realistically.
π
Tamb (t) = 25 + 6 cos (t − 14) (1)
12
t −6
G sun (t) = 965 sin π , 6 < t < 20 (2)
14
1492 D. Chaatouf et al.
Fig. 3 Comparison of the average outlet temperature and velocity with those of [8]
The inlet boundary condition is a pressure type with atmospheric pressure because
the dryer studied works under natural convection, as well as the outlet type. Regarding
the convective losses in the walls a heat transfer coefficient is defined (3)
3 Results
3.1 Validation
To validate our work, we compare our results with certain experimental and theo-
retical results available in the literature as Jyotirmay et al. [8] that studied a cubical
wooden chamber, similar to a cabinet dryer equipped with an inclined solar chimney
for natural ventilation. Our results presented in Fig. 3 shows a good agreement. The
difference rate for temperature and velocity does not exceed 1.47% and 16% respec-
tively. Noting that our results are closer to the experimental results than the numerical
ones calculated by the authors.
3.2 The Optimal Distance Between the Trays and Its Number
In order to obtain an optimum drying chamber, that is capable of carrying out the
drying process with high quality in a short period of time, the optimization of the
dynamic behavior is necessary for the chosen geometry. For the mesh independency
Optimization and Improvement of the Efficiency of a Drying System … 1493
test, four different grids, 5532, 12130, 16606, and 23820 triangular cells were studied,
and the results show that a mesh grid with 12130 cells is quite enough to study the
geometry. We will study the effect of adding trays on the dynamic behavior of the
existing ones. The cases studied are generally two to five trays, and the distance
between the trays is also studied. The objective is to create a high-performance
drying system with an optimal number of trays for high-quality drying. Concerning
the distance between the trays, three different distances were studied between the two
first trays and the results are presented in Fig. 4 in term of the average temperature of
the trays in the sunshine hours, which shows that as long as the distance between the
two trays increases, the temperature decreases along the lower tray, while the first
one remains constant, this is due to the hot air that rises directly when entering the
drying chamber after being heated by the solar air heater.
Furthermore, the Fig. 5 shows the average temperature of the trays in the sunshine
hours for 4 different cases, where the number of trays is the parameter on which the
study is based.
As you can see, the higher the number of trays, the more T between the additional
tray and the first one. The big differences can affect the products, that’s why the
temperature in the trays must be kept at the same level as uniform as possible. It has
been shown in the figure that the temperature in the case of two trays is higher than
the others; and the difference in temperature is very remarkable, after every addition
of a tray compared to the initial case which has an important and significant effect
on the temperature of the existing ones.
In the cases of two and three trays, it can be noticed that the average temperature in
the trays is relatively equal in each case, but when the number of trays is greater than
three, the temperature differences between the plates of each case are getting bigger.
In particular, by decreasing the average temperature in these trays. The difference
can reach more than 10° in some cases, which is not good for the uniformity of
drying and the product will take a longer time to get it dried, especially the product
that was put in the lower trays where the temperature is lower compared to the high
(first) one. In terms of these qualifications (drying uniformity and drying time), the
two-trays case is better than the others, followed by the three-trays case, which is
slightly better than the others, with a temperature difference of around 2°, besides
this, adding a third tray will carry more product, and thus the optimal number of
trays in our case is three trays (Fig. 6).
313
312
311
310
d= 5 cm d= 7,5 cm d= 10 cm
1494 D. Chaatouf et al.
318
316
314
312 Tray 1
310
Tray 2
308
T(K)
306 Tray 3
304
Tray 4
302
300 Tray 5
298
296
5 trays 4 trays 3 trays 2 trays
Fig. 5 The average temperature of trays in the 4 cases in the sunshine hours of the chosen day
V (m/s)
0.00 0.009 0.018 0.027 0.036 0.042 0.048 0.054 0.069 0.075 0.087 0.093 0.1
T(K)
290 296 301 310 316 321 327 333 338 347 353 358 364 370 378 385
(a)
(b)
Fig. 6 Distribution of velocity (a) and temperature (b) inside the chamber at 3 different hours in
the chosen day
The figure above shows the air distribution inside the optimal geometry of the
drying chamber that we studied (the solar dryer was cut off in the pictures). Where
we can see the drying chamber getting hot through the day and the temperature distri-
bution become more uniform after 1 p.m. where the irradiation is at its maximum.
The hot air rises directly when entering the drying chamber after being heated by
the solar air heater and then moves horizontally in the trays before exiting from the
outlet. The air with high velocity passes through the three trays which is not the case
in more than three ones because it crosses just the high trays due to the lower density
of the hot air.
Optimization and Improvement of the Efficiency of a Drying System … 1495
4 Conclusion
In the present work, the optimization of an indirect solar dryer composed of a solar
air heater and a wooden chamber is studied in 2D using ANSYS FLUENT software
under the meteorological conditions of Oujda city (eastern of morocco), the distance
between the trays and its number is a crucial parameter in term of the uniformity of
air inside the drying chamber and according to the results, a drying chamber with 3
trays found to be the optimal number with a distance of 5 cm between the trays.
References
1. Essalhi H, Tadili R, Bargach MN (2017) Conception of a solar air collector for an indirect solar
dryer. Pear drying test. Energy Procedia 141:29–33
2. Goud M, Reddy MVV, Chandramohan VP, Suresh S (2019) A novel indirect solar dryer with
inlet fans powered by solar PV panels: drying kinetics of Capsicum Annum and Abelmoschus
esculentus with dryer performance. Sol Energy 194:871–885
3. Saini V, Tiwari S, Tiwari GN (2017) Environ economic analysis of various types of photovoltaic
technologies integrated with greenhouse solar drying system. J Clean Prod 156:30–40
4. Dhanushkodi S, Wilson VH, Sudhakar K (2017) Mathematical modeling of drying behavior of
cashew in a solar biomass hybrid dryer. Resour-Effic Technol 3(4):359–364
5. Montero I, Blanco J, Miranda T, Rojas S, Celma AR (2010) Design, construction and
performance testing of a solar dryer for agroindustrial by-products. Energy Convers Manag
51(7):1510–1521
6. Chaatouf D, et al (2021) Trays effect on the dynamic and thermal behavior of an indirect solar
dryer using CFD method. In: Hajji B, Mellit A, Marco Tina G, Rabhi A, Launay J, Naimi S (eds)
Proceedings of the 2nd international conference on electronic engineering and renewable energy
systems, ICEERE 2020. Lecture Notes in Electrical Engineering, vol 681. Springer, Singapore.
https://doi.org/10.1007/978-981-15-6259-4_72
7. Amanlou Y, Zomorodian A (2010) Applying CFD for designing a new fruit cabinet dryer. J
Food Eng 101(1):8–15
8. Mathur J, Mathur S (2006) Summer-performance of inclined roof solar chimney for natural
ventilation. Energy Build 38(10):1156–1163
Monitoring and Control System
of a Hybrid Micro-CSP/Biomass Boiler
System for Traditional Baths
Abstract Traditional hammams are big consumers of wood-energy and water. This
work presents a hybrid micro-csp/biomass system for space heating and hot water
production in traditional baths that enables to offset the wood that is inefficient used
in such baths. In addition, energy efficiency of these baths is enhanced through an
optimal control system. Furthermore, the use of local agricultural residues instead
of wood all contribute to reducing the stress on forests. The system to monitor and
control the consumption of hot water and energy is presented. The data can be locally
stored and transferred through internet, together with system faults. Preliminary test
of the micro-csp showed that the maximum temperature achieved at the outlet of
the parabolic trough system was 41.7 °C while at the bottom of the storage tank it
was 34.4 °C corresponding to a 8.7 °C increase even though water was withdrawn
by users. Space heating and hot water production are decoupled which allows to
stabilize the hot water temperature to meet the needs when the boiler is off.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1497
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_136
1498 S. Lamghari et al.
1 Introduction
Traditional baths (hammams) are a part of the Moroccan culture. Actually, there are
about 10 000 hammams in Morocco, including 300 located in the city of Marrakech.
Since, on average, a traditional hammam can consume about 1.5 tons of wood energy
per day, they are regarded as one of the most important consumers of wood energy.
This is mainly a consequence of using the traditional method of heating, where the
water is heated by wood combustion and the energy of the combustion gases is used
to heat the floors of the hammam rooms.
Several studies were performed on traditional baths (hammams) in Egypt, Turkey,
Morocco, Syria and Algeria [1]. They found that three-quarters of the total thermal
energy consumption of a traditional baths was due to water heating. In addition, insu-
lated roofs could reduce the demand for space heating by around 20%. Sibley and
Sibely [2] evaluated the initiatives undertaken in Morocco to reduce wood consump-
tion in hammams, in order to improve their energy efficiency and reduce deforesta-
tion. Sobhy et al. [3], on the other hand, performed a study on an individual hammam
located in Marrakech. Temperatures between 30 °C and 37 °C can be reached for
sunny days in the winter by using a solar heating system without thermal storage.
Most of the studies that target the hammams focused on the different part used in the
installed system. For example, Badran and Hamadan [4] compared the use of solar
collectors and solar ponds as a supply of heat for an underfloor heating system. The
found that the solar collectors are more efficient and break even in shorter time than
solar ponds. For these systems to operate continuously (as in the case of hammam)
and late in the evenings either large systems with storage are required. Another option
is to couple the solar heating systems with another source (ex. Biomass) to remedy
to the problem of the intermittence and to ensure a continuous supply of hot water
and heat, which is the object of this work.
The main objective of the present work is to present a hybrid system, which
consists of parabolic troughs (micro-csp) and biomass boilers that was developed
for heating water and space of a traditional hammam. We focus our attention of the
control and monitoring system. Previous studies have focused on the modeling in
the system in order to evaluate study the energetic performance. It was shown that
there is no significant difference in the temperature of the rooms when copper or
PEX tubing are used in floor heating [5]. In the same aspect there are several studies
on the performance of the underfloor hating [4, 6–10]. An experimental study of the
system, which is the subject of the present work, was carried out to test and improve
the control system used.
Monitoring and Control System of a Hybrid Micro-CSP/Biomass Boiler System ... 1499
The installation consists of six branches of collectors in parallel. Each branch has
four parabolic troughs with a total of 8.24 m collector length and 8.25 m2 -aperture
area yielding a peak power of 6 kW under a DNI of 1000 W/m2 . The total peak
power of the solar field is therefore 36 kW. A single-axis east-west tracking systems
was used in our case.
The boiler used is designed to heat water in a closed loop to a maximum temperature
of 90 °C and uses only biomass feed in granular form. Local agricultural residues
such as argan shells, olive pomace and nuts shell can be used as the biomass feed,
thus allowing to avoid the extensive use of wood energy. The boiler is equipped
with a controller allowing to control the biomass feed rate by adjusting the speed of
rotation of a worm shaft and flow of the air leading to an efficiency as high as 90%.
The role of the monitoring and control is to ensure the operation of the various parts
of system within the specified temperature set points, the protection of personnel and
material, and to help in detecting system faults. It is also designed to shut down a part
of the system without affecting the other parts, which allows to perform preventative
or corrective maintenance without stopping the whole system. A schematic diagram
of the control/monitoring system is shown in Fig. 2:
The main controller uses a simple algorithm. First, the temperatures are measured
through the sensors located in various places of the hammam and the system compo-
nents (boiler, storage reservoirs, rooms, inlet and outlet of the micro-csp plant…).
The measured temperatures are then compared with the corresponding set points
(TCSP, TBoiler…). When difference is higher than the 6 °C for a given sensor, the
corresponding circulating pump is activated until the difference is lower than 4 °C.
Monitoring and Control System of a Hybrid Micro-CSP/Biomass Boiler System ... 1501
The evolution of the outdoor air temperature and irradiance during the testing period
of the hybrid micro-csp/biomass boiler system is given in Fig. 3.
In addition, the temperature of the water at the outlet the micro-csp plant and
the bottom of the solar tank were continuously monitored. Figure 4 shows typical
temperatures profiles at these points. It should be noted that hot water is always
drawn from the storage tanks. At night, the temperature of water in the bottom of the
solar tank is the temperature of feed water from well which is around 25.7 °C. For a
sunny day, the maximum temperature achieved at the outlet of micro-csp plant was
41.7 °C and in the bottom of the tank is 34.4 °C, corresponding to increase by 8.7 °C
due to the fact that water is withdrawal by users from the tanks.
Fig. 4 The evolution of the temperature of water in the bottom of the solar tank and at outlet of
parabolic trough system
For a cloudy day, the maximum temperature achieved was 35.6 °C at the outlet
of micro-csp plant and 28.5 °C in the bottom of the solar tank, corresponding to an
increase of 2.8 °C as result of the water withdrawal by users from the tanks.
The evolution of the temperatures of the biomass boiler and of the two tanks
heated by the boiler is shown in Fig. 5, in addition to the control signals (on/off) of
their respective circulator pumps.
When the temperature is higher than the set-point temperature (60 °C), as clearly
seen for example in the case of the second biomass-heated tank, the corresponding
pump is stopped as it can be seen in Fig. 5a (red curves). However, it can be seen
that the water temperature at the bottom of this tank keeps increasing beyond 60 °C
as result of water coming from the top of the first tank since water is drawn from this
the system.
Figure 5b shows the evolution of the temperatures of the biomass boiler and of
the two tanks heated by the boiler when the underfloor heating system is on. It can
be seen that their respective pumps are stopped when the temperature of the boiler
had dropped below or is within 4 °C of the temperatures of the tank. The pumps are
activated when the temperature of the corresponding tank is less than 60 °C and the
boiler temperature is 6 °C higher this value.
Monitoring and Control System of a Hybrid Micro-CSP/Biomass Boiler System ... 1503
(a) (b)
Fig. 5 Control scenarios of the tanks heated by the biomass boiler (a) first one (b) second one
4 Conclusion
In this work we described a hybrid solar-biomass system for space and water heating
in a traditional Hammam. The focus was made on the system control. A simple
algorithm was used to control the system and regulate the temperatures at various
points. The results show that the control system enabled to avoid the cooling of the
hot water when the water in the biomass boiler loop is colder than the water in storage
tanks when the underfloor heating is activated. This work will allow us to develop
an energy management system that will enable to optimize the energy consumption.
Acknowledgment This study is carried out thanks to the funding from the Solar and New Energy
Research Institute of Morocco (IRESEN) under the InnoTerm/InnoBiomass 2014 program.
References
1. Kristina Orehounig AM (2011) Energy performance of traditional bath buildings. Energy Build
43(9):2442–2448. https://doi.org/10.1016/J.ENBUILD.2011.05.032
1504 S. Lamghari et al.
2. Sibley M, Sibley M (2015) Hybrid transitions: combining biomass and solar energy for water
heating in public bathhouses. Energy Procedia 83:525–532. https://doi.org/10.1016/j.egypro.
2015.12.172
3. Sobhy I, Brakez A, Benhamou B (2017) Energy performance and economic study of a solar
floor heating system for a hammam. Energy Build 141:247–261. https://doi.org/10.1016/j.enb
uild.2017.02.044
4. Badran AA, Hamdan MA (2004) Comparative study for under-floor heating using solar collec-
tors or solar ponds. Appl Energy 77(1):107–117. https://doi.org/10.1016/S0306-2619(03)000
12-6
5. Krarouch M, Hamdi H, Lamghari S, Outzourhit (2018) A simulation of floor heating in a
combined solar-biomass system integrated in a public bathhouse located in Marrakech. In: IOP
conference series: materials science and engineering, vol 353, no 1. https://doi.org/10.1088/
1757-899X/353/1/012005
6. Mokhtari A, Kazeoui H, Boukezzi Y, Achard G (1998) Utilisation d’un Circuit Hydraulique
dans un Plancher pour le Chauffage et le Rafraîchissement des Locaux. Revue Energies
renouvelables 1:17–27
7. Oudrane A, Aour B, Benhamou MHM (2016) Méthodologie pour la détermination de l’ écarte-
ment optimal de la chaîne tubulaire d’ une dalle chauffante. Revue Energies Renouvelables
19:11–19
8. Belkacem N, Loukarfi L, Khelil A, Naji H, Braikia M, Missoum M (2015) Simulation des
charges thermiques dans une habitation pilote à plancher réversible. Nature Technologie
9. Amir M, Lacroix M, Galanis N (1999) Thermal analysis of electric heating floor panels with
daily heat storage. Int J Therm Sci 38(2):121–131. https://doi.org/10.1016/S1290-0729(99)800
49-0
10. El Mays A et al (2017) Using phase change material in under floor heating. Energy Procedia
119:806–811. https://doi.org/10.1016/j.egypro.2017.07.101
Low Capacity Diffusion Absorption
Refrigeration: Experimental
Investigations and Thermal
Characterization
1 Introduction
The DAR, subject of our present investigations, was invented in 1928 by two Swedish
engineers, von Platen and Munters [1]. It is a clean thermally-powered refrigeration
technology that can readily be activated by low- to medium-grade renewable heat. It
has been recognized as one of the most encouraging sustainable technologies for cold
production. The machine’s cycle operates at constant total pressure. It uses ammonia
as refrigerant, water as absorber and hydrogen (or less frequently helium) as a non-
absorbable inert gas, necessary to reduce the partial pressure of the refrigerant in the
evaporator to allow the evaporation process to occur in the uniform pressure device.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1505
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_137
1506 I. Saâfi et al.
The main feature of this machine is its good reliability because it doesn’t have
any moving parts. The circulation of the aqueous ammonia solution is driven by a
bubble pump and that of the gases between absorber and evaporator by gravity.
Numerous theoretical and experimental studies are reported in the literature
dealing with the analysis of the performance of this machine, or part of it, operated
with different energy sources and using various mixtures of working fluids [2–8].
Chen et al. [9] designed a modified generator including a heat exchanger that reuses
the heat released in the rectifier to preheat the rich solution leaving the absorber.
The new cycle configuration showed a significant improvement in cooling coeffi-
cient of performance, COP, up to 50% compared to the original cycle for the same
cooling capacity. Taieb et al. [10, 11] proposed a simulation model for a DAR using
ammonia/water/hydrogen as working fluids. The refrigerator was supposed cooled
with ambient air and activated with solar hot water at 200 °C. An optimal result of
COP of 0.126 associated to a cooling capacity of 22.3 W was found. Freeman et al.
[11] established a thermodynamic model to simulate the performances of a DAR
using ammonia-water-hydrogen as working fluid and integrating solar heat pipes as
a source of heat under the climate conditions of Chennai, India. They found that for a
charge pressure of 14 bar and 200 W heat input, a maximum COP of 0.25 was found
for cooling temperature of 5 ºC.
We report in this paper on our experimental investigations of a low-power commer-
cial refrigerator with a capacity of 25 L, powered by an electrical resistance and
operating according to the von Platen and Munters cycle. The refrigerator uses
ammonia/water as refrigerant/absorbent pair and hydrogen, as pressure equalizer.
All its elements are made of steel.
The objective of the study is to assess the performance of the machine under
various energy input conditions, in steady state as well as in transient operation mode.
A key parameter for the evaluation of the cooling capacity of the refrigerator is the
overall heat exchange coefficient that characterizes the thermal exchange between
cabin and ambient air, i.e. that allows the estimation of heat infiltration from outside in
the isolated cabin. This crucial parameter is generally determined by tests consisting
of electric-resistance heating of the refrigerator cabin in steady-state mode. In the
present study however we deduce this key parameter from a series of unsteady state
temperature measurements during the startup of the refrigerator. To this purpose,
a thermal model of the cabin is developed and a nonlinear regression procedure is
applied to retrieve the overall heat exchange coefficient from the performed transient
tests.
Low Capacity Diffusion Absorption Refrigeration … 1507
2 Experimental Setup
We began our tests by determining the minimum input power required to start the
machine and ensure the production of cold in the cabin. Increasing gradually the
supplied power with each test and monitoring the evolution of the cabin temperature
as well as the temperatures of various locations in the machine, it was found that a
minimum of 22 W heating power is necessary to activate the DAR. For lower energy
inputs (e.g. 20 W) no refrigerant vapor is generated in the boiler and hence, no cold
can be produced, because the solution inside the generator is not hot enough to start
boiling (Figs. 3 and 4). Temperatures of 205 °C and higher are required to initiate
the boiling.
For low heating power in the range 22–35 W, large amplitudes oscillations of the
temperatures (up to 40 °C), particularly in generator (Fig. 3) and evaporator (Fig. 4)
are observed. It was also noticed that by increasing the heating power, the frequency
of the temperature oscillations is augmented and their amplitude reduced. These
oscillations disappear completely for energy input of 43W and larger (Fig. 5).
Fig. 3 Evolution of
evaporator temperature for
increasing heat supply: 20,
22, 24 and 30 W
Fig. 4 Evolution of
generator temperature for
increasing heat supply: 20,
22, 24 and 30 W
1510 I. Saâfi et al.
To evaluate the DAR performance, i.e. the rate of cold production in evaporator Q̇ E V ,
the thermal exchange characteristics of the refrigerator cabin has to be known. In
fact, in stationary mode operation, the cold production in evaporator compensates
exactly the heat leakage Q̇ ext from outside,
Q̇ ext = Q̇ E V = (U A)ext (T AM B − TC AB )
TC AB and T AM B denote the mean temperature inside the cabin and that of ambient air,
respectively. (U A)ext , the product of overall heat transfer coefficient and exchange
surface area, characterizes the thermal exchange between the DAR cabin and its
environment. The knowledge of (U A)ext is thus necessary for the evaluation of the
cooling capacity Q̇ E V and the coefficient of performance of the refrigerator, C O P
Q̇ E V
COP =
Q̇ G
dUcab
= Q̇ ext − Q̇ int (1)
dt
Fig. 5 Evolution of
generator and evaporator
temperatures for heat supply
Q̇ G ≥ 43 W
Low Capacity Diffusion Absorption Refrigeration … 1511
Q̇ ext = Q̇ int = Q̇ E V
dUC AB
= (U A)ext (T AM B − TC AB ) − (U A)int (TC AB − TE V ) (4)
dt
In the case of empty cabin (no water bottles inside), we can write
dUC AB dTC AB
= C p,C AB (5)
dt dt
with C p,C AB denoting the heat capacity of the cabin (hardware + inside air). When
a mass m W of water is added, the internal energy of system is increased, and the lhs
of Eq. (4) is correspondently modified
dTC AB dTW
= α(T AM B − TC AB ) − β(TC AB − TE V ) − γ (7)
dt dt
And in the case of the empty cabin (m W = 0),
dTC AB
= α(T AM B − TC AB ) − β(TC AB − TE V ) (8)
dt
1512 I. Saâfi et al.
Fig. 7 Comparison of
measured and calculated
(Eq. 9) cabin temperature
Now, refrigerator tests are performed in stationary mode and for each heating power
input, the cooling capacity evaluated using (U A)ext . The performances of the inves-
tigated refrigerator are graphically depicted in Figs. 9 and 10. As previously noted,
a minimum heat supply is required to start production of cold by the refrigerator. To
ensure sustainable operation however, a larger activation power, in the order of 36 W,
is needed, producing a cooling capacity of 5 W. Beyond 43 W, the cooling capacity
Q̇ E V stabilizes and the C O P begins to decrease.
As the heat supplied to the generator is progressively increased, the flow rates
of pumped solution and generated vapor in boiler increase too. Raising the flow
rate of the refrigerant vapor in boiler means introducing more liquid ammonia in
the evaporator and hence, larger cooling capacity. Further, an augmented flow rate
of the pumped solution allows absorbing a larger amount of refrigerant vapor in
the absorber, corresponding to an enhancement of the refrigerator performance. This
explains the observed ascending trend of C O P and Q̇ E V in Figs. 9 and 10. Increasing
the heat supply to generator however gives rise to an opposite effect on performance.
1514 I. Saâfi et al.
Fig. 9 Experimental C O P
vs. supplied heat to generator
(steady state operation)
Low Capacity Diffusion Absorption Refrigeration … 1515
In fact, due to the raising temperature in the boiler, more vapor is produced but
containing more steam. The purification of ammonia by condensation of the excess
water vapor in the rectifier implies wasting more of the supplied energy in the boiler
without producing more refrigerant vapor as Fig. 10 shows. Even if more refrigerant
vapor is generated, the low value of the overall heat transfer coefficient (U A)int
limits the heat supply to the evaporator and hence, the evaporation of more liquid
refrigerant. The consequence is the declining refrigerator performance observed in
Fig. 9 and the stagnation of the cooling capacity for Q̇ G larger than 43 W. Additional
heating at the generator beyond the conditions for maximum C O P are not necessary
and should be avoided.
6 Conclusion
This parameter allows the evaluation of the cooling capacity Q̇ E V of the refrigerator
and the coefficient of performance C O P in stable steady-state operation mode.
1516 I. Saâfi et al.
Following optimal performances for ambient air temperatures of 23–25 °C are found
Q̇ E V = 7.5W
C O P = 0.15
References
Abstract The aim of this paper is studied the Elements that affect the circuit
SECE for piezoelectric energy harvesting and the influences on piezoelectric energy
harvesting performance. In addition, the energy harvesting system based on piezo-
electric energy plays an important role in the conversion of vibration energy from
the environment into electrical energy, which can be used by low-power elec-
tronic devices. Concerning the interface circuit, the SECER (Synchronized Electrical
Charge Extraction Regulator) circuit is usually required to rectify the alternating
current (AC) signal into a direct current (DC) signal. In this paper, the properties
of the SECER circuit and the influences on the energy recovery performance are
analyzed and presented by power visualization simulation in the case of low load
and high source frequency, as well as in the case of high load and low frequency.
and whose characteristics can significantly influence the energy harvesting. It can
be seen that the harvesting energy has a close relationship with the characteris-
tics of the SECER circuit components. This study discloses the SECE influences
on piezoelectric energy harvesting performance. In addition, the results show that
the nonlinear SECER technique is 20% more efficient than the standard circuit, in
terms of maximum power and bandwidth, for generators characterized by a moderate
electromechanical coupling coefficient.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1517
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_138
1518 Y. El Hmamsy et al.
The energy extracted per cycle in the SECE technique corresponds to the electrostatic
energy available on the piezoelectric element, which equals (01), with VM being the
voltage at the instant of switching (Fig. 1). Due to the losses in the SECE converter,
the energy recovered by switching is given by (02), with γC the efficiency of the
converter (considered constant) [11].
1
Welectr ostatique max
= C0 VM2 (1)
2
1
(W S EC E )max = γC C0 VM2 (2)
2
The calculation of the voltage VM is done considering that the voltage of the piezo-
electric element is reduced to zero when the displacement is minimum or maximum.
The piezoelectric insert then being in open circuit until the next voltage extremum,
VM is therefore expressed according to relation (3). The expression of the energy
recovered by switching is then given by (4), and we therefore deduce the power
recovered by the SECE, the value of which is given by relation (5). These two quan-
tities are independent of the rectified voltage and the load, and 4γC times greater
than the maximum energy or power recovered with the standard technique.
α
VM = 2 uM (3)
C0
α2 2
W S EC E = 2γC u (4)
C0 M
α2 2
P S EC E = 4f0 γC u (5)
C0 M
Q = α piezo u (6)
dU
I = α piezo = α piezo V (7)
dt
With Q the displaced charge, I the generated current, u the displacement of the
piezoelectric membrane, V the speed of the membrane and α electromechanical
constant [24].
We consider the piezoelectric as an alternating current generator whose ampli-
tude is proportional to the strain rate. The piezoelectric has a significant dielectric
constancy that we take into account by the parallelization of a CPZ capacitor. The
model is therefore the following.
The LT1764 is a low dropout regulator optimized for fast transient response. The
device is capable of supplying 3 A of output current with a dropout voltage of 340 mV.
Operating quiescent current is 1 mA, dropping to <1 μA in shutdown. Quiescent
current is well controlled; it does not rise in dropout as it does with many other
regulators. In addition to fast transient response, the LT1764 has very low output
voltage noise which makes the device ideal for sensitive RF supply applications.
Output voltage range is from 1.21 V to 20 V.
The LT1764 regulators are stable with output capacitors as low as 10 μF. Internal
protection circuitry includes reverse battery protection, current limiting, thermal
limiting and reverse current protection. The device is available in fixed output voltages
of 1.5 V, 1.8 V, 2.5 V, 3.3 V and as an adjustable device with a 1.21 V reference
voltage (Fig. 2).
The LT1764 series are low loss 3 A voltage regulators optimized for fast transient
response. The devices are capable of to supply 3 A at a drop voltage of 340 mV. The
bottom the operating quiescent current (1 mA) drops to less than 1 μA at standstill.
In addition to the low quiescent current.
The LT1764 controllers incorporate several protection functions making them
ideal for use in battery powered systems. The units are protected against the two
inverted inputs. and reverse the output voltages. In backup battery applications where
the output can be maintained by a backup battery when the inlet is pulled down to
earth, the LT1764-X acts as a it has a diode in series with its output and prevents
inversion current flow. In addition, in dual-powered applications where the regulator
load is returned to a negative supply, the output can be pulled under the floor up to
20 V and always let the unit start and run.
1522 Y. El Hmamsy et al.
Fig. 4 Power graph of SECER circuit and STD circuit (Standard) as a function of time
In this part, we will study the SECER circuit making it possible to recover the energy
delivered by the piezoelectric (Fig. 3).
Instead a piezoelectric element we will replace a sine current generator of
maximum current I = 1 mA of frequency F = 105.3 Hz, resistance RL = 600 K
and the filter capacitor is 1 μF [13]. Using the LtSipce software, we simulate the
SECER circuit for 1 s. All used parameters are described in the Table 1 below.
The Fig. 4 shows the evolution of power in terms of time. The power increases
up to a maximum value of 400 μW for 0.5 s. In addition, SECER circuit power is
much greater than STD circuit power during this period.
The Fig. 5 shows the evolution of the power proportionally to a frequency, at during
10 Hz the dissipated power (Pdi) exceeds the output power (Pout), at 50 Hz we notice
the convergence of the powers, and at 100 Hz the capacities become almost equal,
this explains the approach of the resonant frequency.
The Synchronized Electrical Charge Extraction Regulator … 1523
8.0E-02
7.0E-02
5.0E-02
Power (W)
4.0E-02
3.0E-02
2.0E-02
1.0E-02
0.0E+00
10 Hz 50 Hz 100 Hz 150 Hz
Frequency (Hz)
Figure 6 shows the power variation as a function of the load in two cases (50 Hz,
100 Hz), it also shows that the power is almost stable. Therefore, the SECER circuit
improves the power generated in the case of weak structures, the power of the SECER
technique remains almost constant, but the energy extracted and recovered has the
advantage of remaining constant whatever the load.
G (SECER): the gain of a SECER circuit
G (STD): the gain of a standard circuit.
Distribution of the powers recovered and extracted in the case of vibrations of
constant amplitude at the optimum load, normalized against the maximum power in
the standard technique is the SECER technique. Note also that, despite the fact that
the STD circuit has the same parameters of the input of the SECER circuit. The gain
of the STD circuit is lower when the circuit connected to the optimum load (Fig. 7).
However, when the circuits are operating in the 150 Hz frequency range, the STD
circuit gain increases relatively to the SECER circuit gain.
Through the curve it is found that the power changes by changing the frequency and
on the other hand, while the power remains stable despite the change of resistance.
The SECER technique also has the great advantage that it is always adapted whatever
the load resistance, allowing the recovery of non-negligible energy even on relatively
low resistances (Fig. 8).
1524 Y. El Hmamsy et al.
1.2E-03
100 Hz
100 Hz
4.0E-04
0.0E+00
20 Kom 50 Kom 100 Kom 150 Kom 200 Kom
Load (KΩ)
2.8E-06
50 Hz
50 Hz
2.1E-06
Power (W)
1.4E-06
7.0E-07
0.0E+00
20 Kom 50 Kom 100 Kom 150 Kom 200 Kom
Load (KΩ)
20% 19%
18% 16%
15%
16%
14%
12%
10%
7% 7% 7% 8%
8%
6% 5%
4%
2%
0%
10 Hz 50 Hz 100 Hz 150 Hz
G(SECER) G(STD)
1.6E-02
1.4E-02
1.2E-02
Power (w)
1.0E-02
8.0E-03
6.0E-03
1000 Kom
4.0E-03 800 Kom
2.0E-03 600 Kom
0.0E+00 400 Kom
10 Hz 100 Kom
50 Hz
100 Hz
150 Hz
Frequency
Therefore, SECER circuit allows improving the power generated in the case of
weak structures, to improve the bandwidth, and to provide a power theoretically
independent of the electrical load (independent of VDC ).
4 Conclusions
In this paper, a SECER circuit structure is studied and used as interface for piezoelec-
tric system to enhance the performance of the piezoelectric system. The basic working
principles of the piezoelectric structure and several frequently used interfaces, the
influences of resistive load and source frequency,
In the case of SECER, the ripple rate is a priori independent of the parameters
of the active element, but strongly depends on the inductance of the converter. The
harvesting power of SECER is determined by γC , with the intermediate reversal
factor being mainly determined by Cint and L.
The SECER technique presents a better harvesting performance with a load-
independent, the maximum powers are obtained on load resistances distributed
around the optimum load as standard. On the other hand, although exhibiting a
lower power gain, allows energy to be recovered regardless of the load connected to
the micro generator.
In addition, the results show that the SECER non-linear technique is more efficient
than the use of a simple rectifier, in terms of maximum power and bandwidth, for
generators characterized by an electromechanical coupling coefficient.
1526 Y. El Hmamsy et al.
References
1. Zhang Y, Bowen CR, Ghosh SK, Mandal D, Khanbareh H, Arafa M, Wan C (2019) Ferroelectret
materials and devices for energy harvesting applications. Nano Energy 57:118–140
2. Ennawaoui C, Lifi H, Hajjaji A, Elballouti A, Laasri S, Azim A (2019) Mathematical modeling
of mass spring’s system: hybrid speed bumps application for mechanical energy harvesting.
Eng Solid Mechanics 7(1):47–58
3. Ennawaoui C, Lifi H, Hajjaji A, Samuel C, Rguiti M, Touhtouh S, Courtois C (2019) Dielectric
and mechanical optimization properties of porous poly (ethylene-co-vinyl acetate) copolymer
films for pseudo-piezoelectric effect. Polym Eng Sci 59(7):1455–1461
4. Cao Y, Li W, Figueroa J, Wang T, Torres D, Wang C, Sepúlveda N (2018) Impact-activated
programming of electro-mechanical resonators through ferroelectret nanogenerator (FENG)
and vanadium dioxide. Nano Energy 43:278–284
5. Wilson SA, Jourdain RP, Zhang Q, Dorey RA, Bowen CR, Willander M, Johansson C (2007)
New materials for micro-scale sensors and actuators: an engineering review. Mater Sci Eng R
Rep 56(1–6):1–129
6. Malki Z, Ennawaoui C, Hajjaji A, Najihi I, Eljouad M, Boughaleb Y (2020) Pedestrian crossing
system for the mechanical energy harvesting using piezoelectric materials. In: IOP conference
series: materials science and engineering, vol 948, no 1, p 012030. IOP Publishing, November
7. Zhukov S, von Seggern H, Zhang X, Xue Y, Ben Dali O, Pondrom P, Kupnik M (2020)
Microenergy harvesters based on fluorinated ethylene propylene piezotubes. Adv Eng Mater
1901399
8. Hajjaji A, Guyomar D, Pruvost S, Touhtouh S, Yuse K, Boughaleb Y (2010) Tempera-
ture/electric field scaling in Ferroelectrics. Physica B 405(13):2757–2761
9. Rjafallah A, Hajjaji A, Guyomar D, Kandoussi K, Belhora F, Boughaleb Y (2019) Modeling
of polyurethane/lead zirconate titanate composites for vibration energy harvesting. J Compos
Mater 53(5):613–623
10. Hajjaji A, Guyomar D, Touhtouh S, Pruvost S, Boughaleb Y, Rguiti M, Benkhouja K (2010)
Nonlinearity and scaling behavior in a soft lead zirconate titanate piezoceramic. J Appl Phys
108(6):
11. El hmamsy Y, Ennawaoui C, Hajjaji A, Boughaleb Y (2020) Theoretical study and Simulation
method for optimizing the performance of advanced Energy Harvesting techniques. In: IOP
conference series: materials science and engineering, vol 948, no 1, p 012014. IOP Publishing,
November
12. Guyomar D, Lallart M (2011) Recent progress in piezoelectric conversion and energy
harvesting using nonlinear electronic interfaces and issues in small scale implementation.
Micromachines 2(2):274–294
13. Lallart M, Garbuio L, Petit L, Richard C, Guyomar D (2008) Double synchronized switch
harvesting (DSSH): A new energy harvesting scheme for efficient energy extraction. IEEE
Trans Ultrason Ferroelectr Freq Control 55(10):2119–2130
14. Nechibvute A, Chawanda A, Luhanga P, Akande AR (2012) Piezoelectric energy harvesting
using synchronized switching techniques. Int J Eng Technol 2(6):936–946
15. Garbuio L, Lallart M, Guyomar D, Richard C, Audigier D (2009) Mechanical energy harvester
with ultralow threshold rectification based on SSHI nonlinear technique. IEEE Trans Industr
Electron 56(4):1048–1056
16. Erturk A, Inman DJ (2008) Issues in mathematical modeling of piezoelectric energy harvesters.
Smart Mater Struct 17(6):065016
17. Lallart M, Anton SR, Inman DJ (2010) Frequency self-tuning scheme for broadband vibration
energy harvesting. J Intell Mater Syst Struct 21(9):897–906
18. Lallart M, Guyomar D (2008) An optimized self-powered switching circuit for non-linear
energy harvesting with low voltage output. Smart Mater Struct 17(3):035030
19. Ennawaoui C, Hajjaji A, Azim A, Boughaleb Y (2016) Theoretical modeling of power harvested
by piezo-cellular polymers. Mol Cryst Liq Cryst 628(1):49–54
The Synchronized Electrical Charge Extraction Regulator … 1527
Abstract The purpose of this study used image method and superposition theorem
for plot the lateral profile of the magnetic fields around three-phase overhead power
lines based on current return on ground, the analytique results we obtained by
MATLAB numerical software which makes it possible to better analyze and to repre-
sent the 2D transverse profile of the magnetic induction under overhead power lines.
the research works show that strength of the magnetic field was in the middle phase
center of circuit line is very important and decrease rapidly of increase the lateral
distance of conductors. The analysis results can be used to determine the magnetic
field intensity generated by high-voltage power lines and evaluate the factors effects
in the magnetic induction.
1 Introduction
Overhead transmission power lines high voltage, travel long distances and generally
cross regions that are very different in their satisfaction (topography of the land,
altitude, etc.), their climate (temperature, pressure, wind, etc.), their environment
(industrial zones, coastal regions, etc.). As a result, utility of electrical power trans-
mission is exposed to various forms and degrees of stress [1]. The increase in the
human population in the world has resulted the increase demand for the consumption
of electrical energy, which requires the construction of new transmission lines (HV
and VHV) of electricity. The latter has become an essential partner in our daily life,
our technological and economic development [2]. Very low frequency electromag-
netic fields generated by power transmission lines have raised serious questions about
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1529
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_139
1530 S. Houicher and R. Djekidel
the possible biological effects on human health and the environment associated with
the resulting high levels of the intensity of these fields. HV and VHV overhead trans-
mission lines generates magnetic fields which can reach an intensity of 0.4 μT at a
distance of 100 m (400 kV) or at 30 m in (225 kV). The precise evaluation of electric
and magnetic fields produced by high voltage overhead power lines is very important
in several research areas, necessary in many applications in the design, maintenance
and operation of electrical and electronic equipment [3, 4]. There are various exper-
imental and simulation projects studies on electromagnetic characteristics of the
overhead transmission line. The majority of work is based on the assessment and
comparison of the product field under high voltage power lines with international
standards, there are some research studies have also been conducted on decreasing the
field intensity of electric power lines by the technical solutions. A very good modeling
of the circuit lines relies on several parameters, such as: the characteristics of the
electric cables, the height of conductors, the geometric and structural characteristics
of pylons, the electric frequency, the distance between phase conductors and the envi-
ronmental characteristics [5]. Exposure of the magnetic fields intensity generated by
high voltage transmission lines were estimated to raising cancer. For around twenty
years, the environment has been the subject of social concerns, whose exposure
to the electromagnetic fields (EMF) is perhaps the most significant fear expressed
by the human population. From view point of environmental impact, the magnetic
flux density around overhead power line must be measured. It is well understood the
geometric properties of extra high voltage lines (EHV) influences the magnetic induc-
tion and profiles. With the increase power demand, the power transmission utilities
increase the transmission voltage levels for effective magnitude power transmission.
The interference of these low frequency (50 Hz) electromagnetic fields with human
organisms (working personnel and general public) may cause a menace. The subject
of this assumption is discussed in several research projects over the last few decades
[6]. In this work, the MATLAB program for magnetic field calculation was created,
this program was designed to handle various line configurations. For this, a model
including the partitioning of two conductors in each phase, and the effect of catenary
described by the line conductors between towers was considered. Our aim is simu-
lation the lateral distribution of the magnetic field intensity under overhead power
line (400 kV) horizontal configuration in 1 m at the ground level by application
of Garson method, in order show the factors effecting the magnetic induction such
as the loading current, the height of conductors, separation distance between phase
conductors, observed point height above the ground and arrangement of the circuit
line. This paper organizes four parts, the first is introduction. the second defines theo-
rical formulas for calculation of power line magnetic fields based in Garson method
and superposition principle. the third provides the simulation results and discussion;
the fourth is conclusion.
Calculation of the Magnetic Field in the Vicinity … 1531
μ0 .I
B = μ0 .H = (1)
2.π.r
Where: μ0 is the magnetic permeability of vacuum (N/A2).
We can use the theory of conductor images in taking into account the depth of
penetration. The magnetic field produced by the line induces return currents in the
earth, thus creating a magnetic field composed to that generated by the line. Precise
calculation of magnetic induction requires the use of Garson method, which is based
on the notion of return currents through the ground [9]. We consider a fictitious plane
of return currents, parallel to the ground plane and placed at the penetration depth
at a complex distance equal to α (see Fig. 1). Its relation is given by the equation
below:
√
α= 2.δ.e− jπ/4 (2)
√
Where, δ = ρ/μ0 .π. f : depth of penetration, and ρ: the electrical resistivity
of ground (.m), and f: the frequency in Hz. In the simulation, the distribution of
magnetic fields under overhead transmission line is obtained for a balanced system
in which the currents in phase conductors are symmetric and in direct succession at
a phase angle ϕ of 120° from one to the other. the electric current module of the line
is 1000 A, the current of guard cable is neglected and the earth it is assumed that to
be homogeneous with a resistivity of 100 .m. the frequency is f = 50 Hz.
The horizontal and vertical compenents of the magnetic field are calculated by
using the following equations:
⎫
μ0 y − hn y + hn + α ⎪⎪
Bxn = − .In 2
− 2 ⎪
⎬
2.π rcn rin
(3)
μ0 x − dn x − dn ⎪⎪
B yn = .In − ⎪
⎭
2.π 2
rcn 2
rin
Where, (x, y) are the positions of the observation point above ground and (dn , hn )
are the locations of the phase conductors respectively, rcn is the distance between
each conductor and observed point above ground, rin is the distance between each
image conductor and observed point.
The components of magnetic field intensity due to conductor can be written
by adding the two components vertical and horizontal of the given field for each
conductor are given in the expression below:
⎫
N =3
⎪
⎪
Bx = Bxn ⎪
⎪
⎪
⎬
n=1
(4)
N =3
⎪
⎪
⎪
By = B yn ⎪
⎪
⎭
n=1
We consider a three-phase high voltage overhead power line having the arrange-
ment and the geometric coordinates shown in the figure presented below, the hori-
zontal configuration it is a type of circuit lines, the phase conductors are arranged
on the same horizontal plane, that is to say the same height above the ground. it
is usually used for electric power transmission lines. We have three phases, the
horizontal bundle of two conductors contains in each phase are separated by 0.4 m
between them, the radius of conductor is 1.34 cm. According to the coordinates
system presented in Fig. 2, the overhead power line conductors are horizontally posi-
tioned at −10 m, 0 m, +10 m. Conductors vertical separations (heights of conductors)
are 13 m constant between phase conductors and ground, the radius of guard cable
is 0.7 cm positioned at −7 m and +7 m horizontal, and 21 m above the ground,
The span of the circuit line is of 300 m. The schema of overhead transmission line
400 kV used in this study, with the arrangement and geometric is shown in Fig. 2.
Its parameters are presented in Table 1.
For the simulation of magnetic field intensity we found the following results are illus-
trated the lateral profile of highest magnetic induction under three-phase overhead
transmission line. Figures 3(a, b) show the lateral profile of the magnetic induction
1534 S. Houicher and R. Djekidel
Magnetic induction[µT]
overhead transmission lines 10
used by following: a Garson
method, b Biot-savart law.
0
-80 -60 -40 -20 0 20 40 60 80
Lateral distance[m]
(a)
14
Bx
By
12
Bt
Magnetic induction[µT]
10
0
-80 -60 -40 -20 0 20 40 60 80
Lateral distance[m]
(b)
distribution in 1 m from the ground level, On observe in two Figures the lateral
distribution of horizontal, vertical and total components (Bx, By, Bt) of the magnetic
induction, and also the magnetic flux density increases in symmetrically contin-
uous manner for significative decreases of the lateral distance of the high voltage
overhead power line.
For Garson method (see Fig. 3a) the maximum total value directly under middle
phase conductor(about 12.44 μT) register at the center distance(x = 0 m), accounting
for only 12.44% of the exposure limit established by ICNIRP for the general public
100 μT, then reduces rapidly in symmetric way for increase the lateral distance
when one moves away from the conductors, to reach at distance 75 m, far from the
conductors a value is about 0.61 μT almost negligible, On the other hand, For the
comparison results by Biot-savart law (see Fig. 3b) the maximum total value directly
under middle phase conductor (about 12.46 μT) register at the center distance (x =
0 m), accounting for only 12.46% of the exposure limit established by ICNIRP for
the general public 100 μT, then reduces rapidly in symmetric way when one moves
Calculation of the Magnetic Field in the Vicinity … 1535
away from the conductors, to reach at distance 75 m, far from the conductors a value
is about 0.51 μT almost negligible, The calculated magnetic field strength is within
the acceptable range right of way in the corridor high voltage three-phase overhead
power line.
In Fig. 4 plot the lateral profile of magnetic field intensity as a function for different
values of loading current of phase conductors 500, 1000, 1500, and 2000 A. It is clear
that as this curve is symmetric in the two sides positive and negative of the circuit
line, it is show the currents increase, the magnetic induction increase until to reach
a maximum values in all this four values of current in the center distance under
the middle phase conductors, then decrease rapidly with significative increase of the
lateral distance, to reach the low values when one moves away from the conductors,
a very far for the circuit line, this is linearity relationship between them more the
20
15
10
0
-80 -60 -40 -20 0 20 40 60 80
Lateral distance[m]
H of conductor H=18m
H=20m
10
0
-80 -60 -40 -20 0 20 40 60 80
Laterale distance[m]
1536 S. Houicher and R. Djekidel
current is higher, the magnetic induction intense but on observe in the Fig. 5 the
lateral profile of magnetic field intensity as a function for different heights H of the
phase conductors, on illustrate the height H of conductors increases, the magnetic
induction decreases in symmetric manner in the two sides positive and negative of
the circuit line, for significative increase of the lateral distance, to reach the low
values when one moves away from the conductors, a very far for the circuit line,
this is opposite relationship between the conductors height and the magnetic field.
Increasing the line height is the most effective parameter in line design, which is
decrease the field strenght under the high voltage overhead power line.
In Fig. 6 plot the magnetic induction lateral distribution as a function various
observed point heights 0, 1, 1.5, 1.8 and 2 m above the ground, it is clear that as the
observation point heights increase, the magnetic induction increases for the lateral
distance of the circuit line increase in all types of configuration, this is symmetrically
manner linearity relation between them in the right of way of the overhead power line.
the magnetic induction values in the vicinity of single circuit high-voltage power line
depending on the distance to several levels starting with level 0 m, on observe a two
dimensional plot of the magnetic induction lateral distribution for this five levels.
z=1.8m
observation point heights
z=2m
above earth 10
0
-80 -60 -40 -20 0 20 40 60 80
Lateral distance[m]
Figure 7 illustrates the magnetic induction lateral profile in four different phase
configurations: horizontal, vertical, triangular, and inverted triangular circuit line in
1 m from the ground level. By comparing the values obtained, it can be observed that
the magnetic induction values are higher for the horizontal configuration in all points
of the transmission line corridor then decrease rapidly with the distance increase, and
the magnetic induction values for the inverted triangular configuration are lower in
comparison with the magnetic induction values with other configurations (vertical,
triangular). It is very clear that the horizontal circuit line configuration causes the high
values of magnetic induction with a maximum value of 12.98 μT, while the inverted
triangular circuit line configuration obtained the lower values for the magnitude of
the magnetic induction with a maximum value of 6.69 μT in the center distance
of the configuration for the same load current amplitude of 1000 A. This is due to
Calculation of the Magnetic Field in the Vicinity … 1537
Magnetic induction[µT]
configurations of overhead 10
transmission lines
8
0
-80 -60 -40 -20 0 20 40 60 80
Lateral distance[m]
difference in the height of phase conductors above the ground in each circuit line
configuration.
4 Conclusions
In this article used a method for calculation 50 Hz lateral profile of the magnetic
induction produced by high voltage transmission line. This method enables for a
precise evaluation of all currents in the power line, is based on the theory of images
but where the plan of null potential is located at a complex depth, a program for
calculation the magnetic induction was applied. It can handle any line geometry;
accurate results on the 2D quasi-static analysis of the magnetic flux density generated
by EHV transmission line has been obtained. From the simulation results, it is clear
that the magnetic induction is maximum in the center distance of the power line, and
then decreases with increase in the lateral distance of conductors, It has also been
found that the magnetic field depends on several factors, such as the load current,
the height of conductors, the distance between two adjacent phase conductors, the
observation point height, the configuration of the circuit line: horizontal, vertical,
triangular or inverted triangular line, the horizontal configuration creates the high
magnetic induction values due to all conductors being near to the ground level, on
the other hand the inverted triangular configuration produces the lowest magnetic
field values. The magnetic induction values underneath 400 kV overhead power
lines with same amplitude current of 1000 A does not exceed the value 20 μT ,which
is five times lower of the exposure limit established by ICNIRP for the general public
100 μT.
1538 S. Houicher and R. Djekidel
References
Abstract Every year, millions of people around the world die from road accidents,
excessive or unsuitable speed is one of the main factors responsible for accidents, and
it is implicated in a third of deaths. The main objective of this paper is to follow the
driver from the speed of the vehicle as well as to know his behavior and his physical
state on the roads to reduce the risk of accidents. In this paper, we monitor the
instantaneous speed of the vehicle using a control system that can be mounted in any
vehicle equipped with the OBD-2 protocol (on-board diagnostics). A monitoring
chain based on a PIC microcontroller will be mounted in the vehicle as well as
other peripherals such as an SD memory card module, a GPS geographic positioning
sensor, a shock sensor and a global system for mobile communications GSM module.
Unauthorized speed and the location data of the vehicle will be recorded in the SD
memory card with accurate date and clock. In the event of an accident, an emergency
message containing the latest speed and position data will be sent through the GSM
module to the nearby medical center and local police as well as to the operator or
owner (in case of car rental).
1 Introduction
Many electric units are integrated into modern cars as well as relevant technologies
such as the anti-lock braking system (ABS), electronic stability program (ESP) and
the forward collision warning, to improve the safety of driver and passengers, and
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1539
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_140
1540 M. Karrouchi et al.
also to make life easier for the driver. Despite this change, accident statistics world-
wide keep growing due to bad behavior and driver negligence. The major causes of
accidents is unsuitable or excessive speed [1], it is among the most frequent human
causes, is present in more than one fatal accident in 4. This is shown by the analysis
of the causes of fatal accidents, according to the ministry of transport equipment, that
occurred outside the urban area in Morocco, is presented below in Fig. 1. In order
to reverse this tendency, the government is integrating actions having an immediate
effect and added value on improving road safety, It reflects on the installation of
several devices in vehicles, among them a data logger or a black box, it is a device
that facilitates the provision of information on the accident and also records the last
events before the coincidence to know-how was to create the accident.
Several university researchers have proposed different strategies to reduce the
proportions of accidents. In this paper, we have designed an intelligent node dedicated
to the vehicle, this chain behaves like an electrical control unit (ECU), which receives
data of unauthorized or excessive speed data with location, and records them in real
time on an SD memory card as a punctual speed history. In case of a collision, the
latest events of the car will be collected and sent through a telephone message to
emergency numbers using both GPS and GSM modules. Speed data is collected using
the OBD-2 port in the tests, which are performed on a DACIA Dokker. A hardware
and software design has been validated by simulation and practical experience.
The rest of the paper is constructed as follows: the second section recalls the
concepts of CAN bus and the OBD-2 protocol, on which our work is based. The
third section describes some existing techniques for recording vehicle parameters
before the accident. The fourth section presents our proposed system (Hardware and
software). The fifth section details our practical and simulation results to monitor the
driver and the vehicle. Finally, the sixth section concludes this paper.
% of accident causes
40
excessive speed
30
non respect of priority
20
alcohol drink
10
redness
0
Fig. 1 The main causes of fatal accidents in % outside urban areas (according to ministry of
transport equipment)
Implementation of the Vehicle Speed and Location Monitoring System … 1541
2 Relevant Background
CAN (Controller Area Network) bus is a local area network developed firstly for
the field of automotive and industrial automation control. Under the name of multi-
plexing, the CAN is a broadcast serial bus standard for connecting electronic control
units (ECU) peripherals, or microcontroller devices [2]. When a unit fails, it does
not affect others. Each node is able to send and receive messages. Figure 2 shows us
the topology of the CAN module.
As shown in Fig. 1, the standard OBD-2 port is connected to the CAN bus, this
protocol is implemented in most of the vehicles which are manufactured lately. The
OBD-2 standard covers only the exchange of diagnostic data. Figure 3 shows the
OBD-2 connector, The CAN bus pins are pin 6 (CAN high) and pin 14 (CAN low).
The OBD-2 standard defines 10 diagnostic services, the first is used to determine
current information from certain vehicle sensors via the request/response approach
(for example temperature, engine speed, vehicle speed…). The OBD-2 message is
the same useful width as CAN protocol, it supports 8 bytes maximum. The format
of the OBD2 message is shown in Fig. 4.
In mode 01 which defines the request for current diagnostic data, there are several
PID codes defining the data that can be acquired. Table 1 shows some PID codes.
The A and B signifying the fourth and the fifth bytes respectively on the response
frame.
Identifier Number of
Data Data Data Data Not
Request: 7DF additional Mode PID used
A B C D
Response: 7E8 bytes
3 Related Works
Abhirup Das et al. [3] have shown different causes generating fatal road accidents,
effective solutions to fight against them. With the support of various useful sensors,
they react to the speed of the vehicle and define their location. The first step is the
monitoring of the blinking of the driver’s eyes using an infrared sensor, in the event
of sleep, an alarm is triggered. In the second phase, they used an alcohol sensor to
detect the drunken state of the driver. The third step is the use of a proximity sensor
to automatically control the speed of the vehicle, if the distance between the vehicle
and their neighbor lowered below the threshold, the sensor would activate the motor
control circuit and a speed limit occurs. Using a GSM module, the last phase consists
of sending an accident aid message to the nearest station.
Kumar et al. [4] have demonstrated a solution to ensure vehicle safety using IoT
technology. An automatic warning system is implemented for the caution of the
driver, they monitored the performance of the vehicle using various sensors such
as a breath analyzer, an ultrasonic sensor to measure the distance from surrounding
vehicles and an accelerometer. During vigorous consumption of alcohol by the driver,
a message will be sent to the owner or the vehicle authority. In the event of an accident,
the acceleration (tilt) and location data of the vehicle will be sent to the local police
and the hospital.
In this paper [5], Reeja S and Mr. Jayaraj V S have proposed a data recording
system before and during the accident, they integrated a drowsiness detection system
linked to an alarm to wake up the driver, a camera records the vehicle during the road
for analysis in the event of collisions, an automatic light dimmer for night driving so
as not to blind driver on the opposite path, radio frequency identification (RFID) as
well as a GPS and GSM to share the facts for alerting purposes.
Rekha S and Hithaishi B S [6] have realized a monitoring system dedicated
to the vehicle based on several devices such as flame sensor, Temperature sensor,
Accelerometer sensor and Pressure sensor. In the event that one of the magnitudes
Implementation of the Vehicle Speed and Location Monitoring System … 1543
measured by the sensors exceeds the limits, a message containing the exact location
of the car will be sent to the persons concerned using GPS and GSM.
S Sethuraman and S Santhanalakshmi [7] discussed and produced an Internet
of Things (IoT) model of the vehicle black box system, which mainly consists of
the raspberry pi, to communicate between all devices, such as a driver recognition
camera, a gas sensor to detect if someone is intoxicated, pressure sensor, MEMS tilt
sensor (accelerometer) to determine the tilt position of the vehicle. If the threshold
quantity reaches a certain level of any sensor, using GPS and GSM modules, vehicle
location, sensor data and images will be submitted to the owner.
4 Proposed System
In this paper, we have made our own vehicle speed monitoring system, based on
electronic components that are inexpensive and consume less energy. The speed data
is collected using the OBD-2 protocol to validate our tests, which saves us from using
an additional speed sensor and reducing loads. Our solutions and very efficient to
monitor driver behavior on the roads and reduce collisions due to increased speed.
Our system allows highway police, rental car owners and managers to check the
vehicle’s speed history and location. If the speed exceeds the allowable limits, A
message will be sent to the owner contains instantaneous speed data and location of
the vehicle. In the event of an accident, the nearby police and hospital station will be
informed.
The system consists of a PIC18f4580 microcontroller to act and supervise the commu-
nications between the peripherals, a transceiver to adapt the signals on the bus, an
SD memory card for recording speed data, Sound alarm to warn the driver, a GPS
module to locate the vehicle, an shock sensor to detect collisions, a GSM module
to connect to the telephone network. The use of this series of pic microcontroller
has several advantages such as the integration of a CAN controller to manage the
transmission and reception rules, an analog-to-digital converter to translate an analog
magnitude into a digital value. Figure 5 describes the system architecture proposed
in this paper.
The CAN circuit (the PIC and the transceiver) extracts the speed data through the
OBD2 port, and the location data using the GPS module through the UART commu-
nication (Universal Asynchronous Receiver Transmitter), then broadcasts these
messages to the SD memory card, using serial communication SPI (Serial Peripheral
Interface).The shock sensor is used to determine if an accident has occurred during
the speed boost. The transmission rate on the CAN bus is 500 kb/s, and the frequency
value of the quartz connected to the microcontroller used is 20 MHz.
1544 M. Karrouchi et al.
Our program implemented in the microcontroller can be presented in two phases, the
first contains a program whose role is to collect the speed data exceeding the defined
limits, to display them on the LCD screen as well as the spatial coordinates of the
vehicle (Latitude and longitude). This information will be recorded in real-time on
the SD memory card as well as sent to the owner or the rental manager. The second
task is to check if a collision has occurred. In this case, an emergency message will
be broadcast to the emergency number, to the nearby police and hospital stations.
The program was written in the MicroC pro for Pic compiler and loaded into the
PIC using the WinPic software. Figure 6 summarizes the flowchart followed.
When the vehicle battery turns on, using a voltage regulator (12 V to 5 V) inte-
grated in the electronic circuit of the transceiver MCP255, our main development
board electronic managed by the microcontroller starts to work. The LEDs shows the
state of the voltage generated on the board. Printed circuit boards are designed using
Altium Designer software. Figure 7 describes the 3D design of electronic develop-
ment boards. After several tests were carried out to ensure the correct functioning of
the system, we were able to determine and display the instantaneous speed data of
the vehicle. Figure 8 shows the materials and the work done in this step.
According to our system and using an oscilloscope, we were able to extract
and visualize the electrical signals defining the CAN transmission frame. It mainly
consists of 11-bit identifier (ID = 0x7DF), 64-bit data field, contains mode 01,
PID 0x0C and data (A, B, C, D/F: not used). Figure 9 shows the electrical signals
characterize the speed data request message.
Implementation of the Vehicle Speed and Location Monitoring System … 1545
a b
Due to the collisions of the many frames circulating on the CAN bus, it is hard to
collect the desirable reception frame, relating to the speed data using the oscilloscope,
since the latter is connected directly to the physical bus of the vehicle.
After this phase, the processing of comparison between the speed data and the
defined limits begins in the microcontroller. The second step is based on a simulation
of the system, which contains the SD card, the GSM and GPS modules connected to
the PIC18F4580 microcontroller on Proteus. It is a software that allows the design
of a virtual electronic system and to simulate it, designed for industrial and educa-
tional use [8]. Proteus integrates the simulation of animated components, circuits and
1546 M. Karrouchi et al.
Fig. 9 Capture of data transmission signals on the vehicle’s CAN bus using an oscilloscope
Fig. 10 The fundamental scheme of the system and the result of the simulation in ISIS Proteus
we noticed that it continually updates and works fine with the change of data. The
excessive speed data taken as an example is 100 km/h.
Exaggerated speed is one of the most common causes of road accidents, this factor
concerns the driver and his behavior. This paper focuses on the design of a surveil-
lance system that determines where the speed of the vehicle is enormous. We used
a hardware platform composed of a PIC18f4580 microcontroller and anmcp2551
transceiver for CAN communication with car, a 12 V to 5 V voltage regulator to
power the system from the vehicle battery, a GPS module to locate the car, a memory
card that records speed and location data, a crash sensor to determine if a collision
1548 M. Karrouchi et al.
has occurred, GSM module to send an important message to the people concerned, a
sound alert to alarm the driver in the event of excitement and an LCD to display the
results. The whole control system has high reliability and can translate into a small
volume. An efficient solution to monitor and know the behavior of the driver, as well
as minimize the risk of accident.
Our work can be improved by the implementation of several devices such as
an alcohol sensor, vibration sensor, we also want to realize a dialogue between the
present control system and a system which monitors the drowsiness of the driver
using a camera device.
References
Abstract Inductive power transfer (IPT) systems are designed to efficiently deliver
power from a primary source to one or more secondary loads via magnetic coupling
over relatively large air gaps. The traditional IPT system is improved by adding a
compensation system to compensate for the reactive power due to leakage fluxes
between the coils, thus improving the system efficiency and increasing the transfer
distance. There are different compensation topologies, but the two most suitable for
the electric vehicle charging system are the SS and SP topology. In this paper, a
comparison between Series-Series and Series-Parallel topology is conducted. This
comparison will help to determine the topology that will ensure maximum transfer
efficiency even over long distances. Simulation results are given to demonstrate the
best topology to improve transfer.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1549
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_141
1550 N. Oumidou et al.
1 Introduction
Growing environmental concerns such as global warming and air pollution make
it necessary to find a solution to the problem of greenhouse gas (GHG) emissions,
which is worsening the situation. So as one of the major consumers of fossil fuels,
the transportation sector is responsible for a large part of these emissions. Hence
the need to minimize this contribution by opting for electric vehicles as a solution
because of the advantages they offer in terms of performance, emissions, and safety.
However, the acceptability of EVs is linked to their autonomy, purchase cost, avail-
ability, and performance of the charging infrastructure. The objective is, therefore,
to improve EV autonomy without having to increase battery volume, and to have a
high-performance, available recharging infrastructure. It is in this context that the
idea of using wireless recharging as a solution to achieve the desired objectives was
born.
Wireless charging is one of the three methods proposed as alternatives for charging
electric vehicles in addition to battery exchange and conductive charging. The battery
exchange allows us to replace the empty battery with a full one in a fast way and
allows a flexible charging time, but it only minimizes the time. As for the conductive
charge that is currently used, it is cost-effective and feasible but has the disadvantage
of a long charge time and also shows some safety problems related to the use of the
cables.
For the wireless charge, it is in the spotlight thanks to its advantages over the
conductive charge. It is one of the five key technologies mentioned in [1] that will
enable a transition to electric mobility and will disrupt the transportation industry
while eliminating the need for cables and the problems of independence anxiety.
Authors in [1] also highlights the importance of integrating sustainable electricity
sources such as solar and wind power into the charging process to make electric
vehicles truly green.
The wireless power transfer technology (WPT) is then a feasible solution that
will boost the range of the vehicle while minimizing the volume of the battery and
consequently the cost. There are different vectors for wireless energy transfer which
are classified as far-field WPT and near-field WPT as in [2]. For far-field WPT it
allows the transfer of power by electromagnetic radiation (laser, microwave, and
radio waves), it is a technique that allows the transfer of high power over long
distances. While WPT in the near field allows power transfer over short distances
using inductive coupling or capacitive coupling.
The two most common wireless transfer techniques are inductive transfer (IPT)
and capacitive transfer (CPT), which have common applications such as charging
some mobile devices and low-power biomedical devices [3].
Currently, this capacitive coupling is used to transfer low powers (a few tens of
watts) over small distances. However, work is underway to improve this technique
and extend it to the recharging of electric vehicles [4]. So, for the EV application
inductive transfer prevails. It is currently the most mature technology for contactless
recharging of electric vehicles. It allows the transfer of several tens of kilowatts over
Comparison Study of the Resonant Inductive Power Transfer … 1551
short distances (a few tens of centimeters) with good efficiency and without danger to
the human body, if a correctly dimensioned magnetic shield is present. Technique is
attractive for EV applications because of its advantages in terms of maturity, transfer
capacity, transfer distance and security. Some of the advantages of IPT are listed
below [5, 6]:
• The simplicity of connection thanks to the absence of cables to handle.
• Safe and reliable system.
• Low maintenance compared to the current system (conductive load).
• Safe load whatever the climatic conditions (system unaffected by dirt, rain, snow,
etc.).
• System with long product life.
But despite these benefits, it should be noted that there are some health and safety
concerns regarding human exposure to EMF and various research is being conducted
to find exposure limits [7]. Researchers in [8] indicates that the two most common
exposure limits are those published by the International Commission on Non-Ionizing
Radiation Protection (ICNIRPT) and the Institute of Electrical and Electronics Engi-
neers (IEEE). Also, the Ref [1] mention that a standard is under development as part
of SAE-J2594, this standard aims to standardize a small frequency range for light
EV load.
In an IPT system, there are two electrically insulated sides; the transmitter side
called the primary side which is usually fixed or integrated to the ground and the
receiver side called the secondary side located under the vehicle. The power transfer
is achieved via magnetic induction between these two components and the charging
process begins when the transmitter coil and receiver coil are fully or partially aligned.
It can be used in static, dynamic, or quasi-dynamic recharging mode. In the case
of stationary or static charging (SWC), the vehicle is charged while stationary during
long-term parking.
The Dynamic Load Control (DWC) allows the vehicle to be loaded while in
motion. In this mode, a set of primary coils follow one another to form a “continuous
strip” on a track or a portion of the road that will continuously power an electric
vehicle while it is moving. Figure 1 shows an example of one of the commercially
available applications of the DWC; the On-Line EV (OLEV) developed at the Korea
Advanced Institute of Science and Technology (KAIST) discussed in [9], which
confirms that the DWC system is significantly beneficial for battery viewing time
and can reduce the size of the battery, but points out that the battery could suffer
serious degradation if it is too small.
Another model of the DWC system is presented in [10], but this time with a single
longer transmission coil. This system will allow loading several VEs at the same time.
This mode, also called on-road power supply, is the future-oriented development for
charging electric vehicles and will allow increasing the number of EVs [1]. Different
possible architectures for the DWC can also be found in [11].
For the quasi-dynamic charging mode (QWC) it is beneficial in the case of tran-
sient stops where vehicles stop at regular intervals such as traffic lights, cab ranks or
bus stops [2].
In this paper, the main contribution is to make a comparison between serial-
to-serial and serial-to-parallel topology in order to choose the best compensation
topology. The comparison consists in varying the frequency and the gap and noting
the evolution of the efficiency of the system and then comparing the two systems to
determine the topology that ensures a better transfer efficiency.
The rest of this paper is organized as follows. Section 2 describes the resonant
system and the inductive coupler as a key element of the system and mentions the two
parameters influencing the transfer efficiency. Section 3 presents the two topologies
that are the subject of the comparison and gives expressions of the efficiencies of
each system found from equivalent electrical circuits. Results and their discussion
are provided in Sect. 4. Finally, Sect. 5 ends this paper with conclusions.
2 System Description
(a) (b)
Fig. 2 Illustration of the traditional inductive power system a [5], resonant inductive power system
b [12] for electric vehicle
Comparison Study of the Resonant Inductive Power Transfer … 1553
1 1
Cs ω0 2 L s ω0 2 L s 1 1
2 2
ω0 2 L s − M
Lp ω0 2 L s − M
Lp
Most of the work cites the four basic compensation circuits as follows: Series-
Series (SS), Series-Parallel (SP), Parallel-Parallel (PP) and Parallel-Series (PS) which
are illustrated in Fig. 3 with the relationships for calculating the primary capacity Cp
and the secondary capacity Cs in Table 1 [7, 8, 13–16].
However, the Refs [7, 8] and [17] present other compensation systems or networks,
namely: Serial-parallel combination, LCL compensation and LCC compensation.
The authors of [18] present an analysis of another topology: the topology of the
double resonant LCCL network.
In what follows, we will content ourselves with our study of the two topologies;
the SS and SP topology, which are more suitable for variable load conditions because
resonance is guaranteed (according to the table of different topologies) [2].
In an IPT system, the inductive coupler constitutes the key element of the wireless
transfer system. It is composed of transmitter coil and receiver coil separated by an
air gap. It is a transformer characterized by the absence of the magnetic circuit;
air transformer. In addition to the two coils a ferromagnetic shielding consisting of
ferrite and aluminum is used in the design of the coupler. This shielding protects
passengers from the risks associated with exposure to magnetic radiation (Fig. 4).
Several geometries can be used to form the inductive coupler. The basic geome-
tries described in the literature use circular or rectangular coils. However, in order
to improve the coupling between primary and secondary, various authors propose
combinations of several windings at the primary and secondary levels. For the wires
used in the design of these geometries, and in order to reduce losses in the windings,
multi-stranded yarns and more precisely Litz wire are used.
1554 N. Oumidou et al.
(1)
With µ0 is the permeability of the vacuum, dl1 and dl2 are respectively the sections
of circuits C1 and C2 and r12 is the distance between sections dl1 and dl2.
In the case of two windings of N1 and N2 turns, the mutual inductance between
the two circuits is written:
μ0 πr1 2
M1−2 = N1 N2 . r2 2 (2)
2 r 2 + Z 2 23
1
μ0 πr2 2
M2−1 = N1 N2 . r1 2 (3)
2 r 2 + Z 2 23
2
Where r1 and r2 represent the winding radius and Z the distance between the two
windings (air gap).
It is related to the coupling factor by the following relationship:
M M
K = =√ (4)
Mmax L1 L2
Figures 5 and 6 show the influence of the increase of the air gap between the primary
and secondary windings on the mutual and the coupling factor.
According to the curves in Figs. 5 and 6, it is clear that the closer the two windings
are to each other, the higher the coupling factor, i.e. the lower the flux losses and the
higher the efficiency of the system. Also, it can be seen that the greater the transfer
distance, the lower the mutual inductance M, which also influences the efficiency of
the system.
Comparison Study of the Resonant Inductive Power Transfer … 1555
The various works on the WPT for EV have mentioned that the addition of a compen-
sation system improves the efficiency of the system as well as the transfer distance.
In this section we will compare the two so-called adequate compensation topologies;
the SS and SP topology shown in Fig. 7.
From the equivalent circuit of the system Fig. 7 and as in [20], the efficiency of
the system in SS structure is
RL
η SS = 2 (5)
[R2 + R L ] + (ωM)
R1
2 (R L + R 2 ) 2
+ L 2ω − 1
C2 ω
(a) (b)
Fig. 7 (a) WPT system with SS compensation and (b) WPT system with SP compensation
RL
ηS P =
2
R1 2
R L + R2 (1 + (R L C2 ω) ) +
2 R L + R2 − R L C2 L 2 ω2 + ω2 (L 2 + R L C2 R2 )
(ωM)2
(6)
In this comparison we will first vary the frequency for a fixed air gap and then we will
vary the air gap for a fixed frequency and we will note the influence of this variation
on the efficiency. Parameters values used in this comparison are shown in Table 2.
With L 2 represent the self-inductance, R1 and R2 the wire resistance and R L the
load.
In this section, the simulation results are conducted on MATLAB software to show
the variation of efficiency as a function of frequency and air gap. The curves obtained
are shown in Figs. 8 and 9.
According to the curves obtained, the more the transfer distance increases, the
lower the yield. Also, it is noted that the SP topology allows to have a good efficiency
for high air gaps 80% at 40 cm while for the same distance the efficiency of a system
with SS compensation is 46%.
Regarding the influence of the frequency variation, we notice that for the system
with SP compensation by increasing the frequency the efficiency of the system
increases until it reaches a maximum value of 83% for a frequency of 45 kHz and
then begins to decrease.
Comparison Study of the Resonant Inductive Power Transfer … 1557
While for the system with SS compensation the efficiency increases by increasing
the frequency up to a maximum value of 94% and then becomes stable.
Then we can conclude that according to the results obtained theoretically the SP
topology is the most adequate. This result remains to be verified experimentally.
5 Conclusion
transfer efficiency but the problem of leakage flux between the coil influences the
transfer quality. So, to minimize these leaks, a compensation system is used in the
form of series and/or parallel connection of capacitors with the winding. Different
topologies can provide the compensation but there are four simple basic topolo-
gies, two of which prevail over the others, the SS and SP topology. The theoretical
comparison between the two has shown that the SP topology ensures good transfer
efficiency even for large distances. This result remains to be confirmed or refuted
later by experimentation.
References
1. Mouli GRC, Venugopal P (2017) Future of electric vehicle charging. In: 19th international
symposium power electronics Ee 2017, Novi Sad, Serbia, 19–21 October 2017
2. Ahmad A (2018) A comprehensive review of wireless charging technologies for electric
vehicles. IEEE Trans Transp Electrif 4(1). https://doi.org/10.1109/TTE.2017.2771619
3. Dai J (2015) A survey of wireless power transfer and a critical comparison of inductive and
capacitive coupling for small gap applications. IEEE Trans Power Electr 30(11). https://doi.
org/10.1109/TPEL.2015.2415253
4. Li L (2020) A family of compensation topologies for capacitive power transfer converters for
wireless electric vehicle charger. Appl Energy 260:114156
5. Ferrieux J-P (2018) Conception d’un système de transfert d’énergie sans contact bidirectionnel
par induction. In: Symposium De Genie Electrique (SGE 2018), Nancy, France, 3–5 Juillet 2018
6. Khutwad SR, Gaur S (2016) Wireless charging system for electric vehicle. In: International
conference on signal processing, communication, power and embedded system (SCOPES)-
2016
7. Patil D (2018) Wireless power transfer for vehicular applications: overview and challenges.
IEEE Trans Transp Electrif 4(1). https://doi.org/10.1109/TTE.2017.2780627
8. Bi Z (2016) A review of wireless power transfer for electric vehicles: prospects to enhance
sustainable mobility. Appl Energy 179:413–425
9. Jeong S (2019) Charging automation for electric vehicles: is a smaller battery good for the
wireless charging electric vehicles? IEEE Trans Autom Sci Eng 16(1). https://doi.org/10.1109/
TASE.2018.2827954
10. Tan L, Guo J Efficiency optimization of on-road charging for electrical vehicles. In: 7th Asia
pacific international symposium on electromagnetic compatibility
11. Gori P-A, Sadarnac D (2017) Sensorless inductive power transfer system for electric vehicles:
strategy and control for automatic dynamic operation. In: 19th European conference on power
electronics and applications (EPE 2017 ECCE Europe)
12. Kalwar KA (2015) Inductively coupled power transfer (ICPT) for electric vehicle charging – a
review. Renew Sustain Energy Rev 47:462–475
13. Aziz AFA Review of inductively coupled power transfer for electric vehicle charging. IET
Power Electr
14. Hammoud A (2017) Charge par induction de véhicules électriques : Analyse du potentiel,
limitations, nouveaux concepts. Le 24 novembre 2017
15. Panchal C (2018) Review of static and dynamic wireless electric vehicle charging system. Eng
Sci Technol Int J 21:922–937
16. Machura P (2019) A critical review on wireless charging for electric vehicles. Renew Sustain
Energy Rev 104:209–234
17. Al-Saadi M, Ibrahim A Analysis and comparison of resonance topologies in 6.6kW induc-
tive wireless charging for electric vehicles batteries. In: The 12th international conference
interdisciplinarity in engineering
Comparison Study of the Resonant Inductive Power Transfer … 1559
18. Jin X (2019) Calculation and analysis of optimal design for wireless power transfer. Comput
Electr Eng 80:106470
19. Raju S (2014) Modeling of mutual coupling between planar inductors in wireless power
applications 29(1).https://doi.org/10.1109/TPEL.2013.2253334
20. Hasanzadeh S (2013) Efficiency analysis of contactless electrical power transmission systems.
Energy Conv Manage 65:487–496
Magnetic Chargers in Electrical Models:
Operating Principle and Efficiency
Analysis of an Inductively Coupled
Power Transfer System
Abstract Wireless charging can automate the charging of electrical devices due to its
simplicity operation, robust, and safety. The basic principle of wireless power transfer
(WPT) system is based on the induction coupling between the primary (transmitter)
and secondary (receiver) coils. This paper proposes and describes a new efficient
wireless recharging system which will maximize the transfer of energy to the usage
part. Initially, define the operating principle and the way to charge electric vehicle
(EV). (i) two electric recharging models are described and studied; (ii) the equations
and the efficiency of each recharging model are quoted and discussed. On the other
hand, a typical WPT system consists of a circular transmitter coil, two capacitors
for leakage compensation, one in the primary side and one in the secondary side
connected either in series or in parallel has been constructed in this work. Finally,
several parameters have an effect on the energy transfer efficiency such as the value
of the added capacitors, the value of the operating frequency, the choice of the
compensation topology (serie-serie: serie-parallel: parallel-serie: parallel-parallel)
are discussed in this paper.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1561
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_142
1562 A. Elkhatiri et al.
1 Introduction
During the last few years, the electrification sector of transport, such as tramways,
trains and electric vehicles, has become one of the most attractive areas of develop-
ment. In a special way the electric vehicle is attracting more and more attention of
the industrialists because of economic and environmental aspects. [1] the use of the
connecting cable for recharging electric vehicles (VE) presents risks for the user, to
solve these problems, the EV can be recharged by using a wireless power transfer
system, in particular the inductive coupler power transfer mechanism (ICPT). In the
existing literature, some approaches have has been proposed for Rechargeable Elec-
tric Vehicles. The implementation of a wireless charging system for electric vehicles
based on inductively coupled power transfer technology, to transfer electrical energy
from the source to the vehicle battery in order to solve the problems of EV charging
[2]. In the paper [3], the authors compare the shapes of the power coils with the intro-
duction of a ferrite core to design an efficient power supply for wireless charging of
EVs. Resonance frequency is a key factor for improving transfer efficiency in a Wire-
less Power Transfer (WPT) system which has a wide application because the wired
media is eliminated. In this paper, an adaptive frequency measurement method is
presented and the influence on the transfer efficiency in the WPT system is explored
in [4]. In this paper [5], the magnetic equivalent circuits and analytical models of
non-contact electrical power transfer systems are developed and evaluated through
experiments. Overall coupling the coefficient and overall efficiency are introduced as
a means of evaluating the performance of the systems. Compensation of capacitors
on the primary and secondary sides of the systems improves the overall coupling
coefficient and overall efficiency. [6] In this paper, a new method of impedance
matching and power distribution using impedance inverters only on the receiver side
is proposed. The mathematical equations of the proposed method are then generalized
for an arbitrary number of receivers and an arbitrary number of repeaters. This paper
[7] presents a state-of-the-art literature review on recent advances in load coupling
design. This document [8] presents the design and implementation of a wireless
energy transfer system for the movement of electric vehicles, as well as an example
of an On-Line Electric Vehicle (OLEV) system. Electric vehicles are charged on
the road using wireless energy transfer technology. The interoperability between the
two coil topologies is studied. The proposed transformer structures with aluminum
shielding meet the human exposure regulations of the International Commission on
Non-Ionizing Radiation Protection (ICNIRP) guidelines as a basis. Finally, exper-
iments have been carried out in [9] to validate the analyses. In [10], an overview
of a typical WPT system has been given. Simple design equations were given to
calculate inductance, capacitance, power, quality factor and coupling coefficient to
optimize the coil design for the electric vehicle application. In addition, a comparison
was made between the most common coil shapes and the effects of changing coil
parameters such as number of turns, pitch and inner and outer radius on coil effi-
ciency were studied. This paper presents a study on wireless charging technology.
The absence of the magnetic core in the magnetic coupler and the transfer operates
Magnetic Chargers in Electrical Models … 1563
with high air gap values, causing leakage and losses that will minimize the system’s
performance. For this reason, we will consider adding two capacitors, one in the
primary part and the other in the secondary part, and increase the frequency of the
system. There are four types of compensation (serie-serie, serie-parallel, parallel-
serie, parallel-parallel). The objective of this study is to select the most appropriate
type of compensation for the wireless charging system by setting the value of the
operating frequency to the frequency of the secondary resonance system and finding
the value of the primary capacitance to minimize the apparent input power and mini-
mize the reactive power of the power supply to maximize the transfer of energy to
the vehicle battery. The paper is organized as follows: Sect. 2 presents an overview
of the principle of wireless recharging. The evaluation of the efficiency of a tradi-
tional magnetic coupler is discussed in Sect. 3. Description of a resonant magnetic
coupler recharging system, and the improvement of their transfer efficiency is given
in Sect. 4. Section 5 discusses this work and the end of it is concluded in the last
section.
The electric vehicle essentially consists of a battery system that acts as a fuel reservoir
and an electric motor. Recharging the battery of the electric car is done according to
the following procedures [11]: Charging station on the road, near shopping centers or
car parks. A recharging cable is necessary which makes the operation tedious since
the recharging is done in a repetitive way and the cable is always in contact with
the ground and can contribute to get your hands dirty and easy to forget in case of
emergency. So, induction recharging came as a solution to these problems. That is,
we’re going to use a wireless charger to charge the vehicle. Wireless charging allows
electric cars to be charged without any physical contact between the primary coil
located on the ground and a secondary coil located above the vehicle chassis. This
solution offers ease of use and good robustness. A general diagram for a wireless
magnetic coupler is shown in (Fig. 1).
A traditional magnetic coupler is a wireless charging system that works with two
electrical coils, placed at a distance called the air gap (Fig. 2):
The electric model of a traditional system:
with: L 1 , L 2 are respectively the primary inductance and the secondary inductance.
R1 , R2 are the primary wire resistance, the secondary wire resistance respectively
and M is the mutual inductance between the primary and secondary circuits.
The expression of the mutual inductance M12 for two circular circuits will be:
M 12 = N 2 N 1 μ20
R12
3 π R22 (3)
2
R12 +Z 21
with: N2 = 7 spires are the Number of secondary windings, N1 = 7 spires are the
Number of primary windings, μ0 = 4 π * 10−7 H.m−1 is the Vacuum permeability,
R1 = 40 cm et R2 = 40 cm are respectively the radius of the primary spiral and the
secondary spiral. Hence the expression of mutual inductance becomes:
N 2 = N 1 = N : R2 = R1 = R : Z1 = Z2 = Z (4)
M 12 = M 21 = N 2 μ20 R4
3 π (5)
(R 2
+Z 2 ) 2
M 12 = M 21 = N 2 μ20 Rπ (6)
But, physically speaking, when the two windings coincide, almost all the field lines
emitted by the first conductor pass through the second conductor. This means that,
theoretically, the value of the mutual inductance tends towards the value of the proper
inductance of the coil (Fig. 3).
μ0
L1 = L2 = N 2 Rπ = 38, 68 μH
2
the mutual decreases rapidly if the air gap in the system increases. Therefore, the
magnetic coupling weakens when the distance d exceeds 1 m.
V 1 = R1 I 1 + j wL 1 I 1 − j w M I 2 (7)
0 = (R L + R2 )I 2 + j wL 2 I 2 − j w M I 1 (8)
Hence:
ηt = RL
(9)
R1
w2 M 2
((R2 +R L )2 +(L 2 w)2 )+R2 +R L
1566 A. Elkhatiri et al.
A resonant charging system is one of the most well-known and advanced versions
of a traditional charging system, in terms of power electronics design and wireless
transformer coils. In addition, compensation networks are added to the primary and
secondary windings not only to create the resonance phenomenon but also to mini-
mize additional losses. Increasing the gap between the two coils and the absence of
a magnetic core increases leakage and voltage drops in the magnetic coupler. As a
result, the magnetic field generated has a negative effect on the mutual inductance
which decreases the value of the magnetic coupling coefficient. (k) the expression
of k is formulated as follows:
K = √M
L1 L2
(10)
It can be seen that the value of k falls rapidly as a function of the air gap d. Therefore,
if we minimize flux losses, the two circuits (primary and secondary) will be strongly
coupled, the mutual inductance will be higher and vice versa.
Magnetic Chargers in Electrical Models … 1567
Since the two circuits are separated by a large air gap which results in a low mutual
inductance M. There are four compensation topologies which are illustrated SS SP PS
PP [12]. The reflected impedance Z r can be found by dividing the reflected voltage
by the primary input current [13]:
w2 M 2
Zr = Zs
(11)
Z s = j wL s + 1
j wC s
+R (12)
For series compensation, the reflexive resistance and reflexive reactance are:
w 4 M 2 C 2S R w3 M 2 C s (1−w 2 L s C s )
Zr = + j (13)
(wC s R)2 +( 1−w 2 L sCs )
2
(wC s R)2 +(1−w 2 L s C s )
2
Z s = j wL s + 1
j wC s + R1 (14)
Similarly:
w2 M 2 R w 3 M 2 [ C s R2 (1−w 2 L s C s )−L s ]
Zr = + j (15)
R (1−w 2 L s C s ) +(wL s )2 R2 (1−w 2 L s C s ) +(wL s )2
2 2 2
Zt = 1
j wC p
+ j wL p + Z r (16)
Yt = 1
Zt
= j wC p + 1
j wL p +Z r (17)
In the procedure for the design of a resonant Inductively Coupled Charging System,
he excites two fundamental hypotheses:—The operating frequency equals the deter-
mined nominal resonance frequency. f 0 = 2π √1C L = √1 —The input current
s s 2π CpLp
is constant. With a changing load there will be a shift in operating frequency So, the
point of secondary compensation is to increase the power transfer capacity. It is
1568 A. Elkhatiri et al.
necessary for the system to operate at the resonance frequency of the compensated
secondary system in series or parallel. The value of the resonance frequency is:
f 0 = 2π √1C L So, to minimize the apparent input power and minimize the reactive
s s
power of the power supply when the system is operated at the secondary resonance
frequency, it is necessary to add a capacitance in the primary circuit by designing the
imaginary part of Z t or Yt equal to zero at the secondary resonance frequency. The
solutions are given in Table 1 for the four topologies.
Depends on: Selected primary and secondary topologies, magnetic coupling coeffi-
cient k and secondary quality Qs (Fig. 5 and Table 2):
The results obtained show that: SS topology is the best topology because the
normalized primary capacitance is independent of the magnetic coupler or secondary
quality Qs.
Table 2 The expression of the normalized primary capacitance for the four topologies for different
values of k
Magnetic Chargers in Electrical Models … 1569
RL
ηss = [15] (21)
R1
w2 M 2
((R L + R2 ) + L 2 w −
2 1 2
C2w
) + (R2 + R L )
1570 A. Elkhatiri et al.
with: W = 2πf: with f the system frequency. C2 = 0.31 uF: secondary capacity. The
operating frequency f varies between: [3000–1000000] Hertz.
According to the curve, the efficiency of the system increases with increasing the
frequency: At 261 kHz, for M = 1.97e−05, an air-gap d = 30 cm: the value of ηss
is of the order of 94%. Above the 261 kHz the value of ηss remains constant.
[15] ηsp= RL
(22)
[R L +R2 (1+( R L C 2 w)2 )]+ w2 M 2 [( R L +R 2 −R L C 2 L 2 w ) +w (L 2 +R L C 2 R 2 ) }
2 2
R1 2 2
5 Discussion
The principal conditions for delivering maximum power are: 1-Operation at the
secondary resonant frequency. 2-Deduct the values of the added primary capacitors.
3-Choosing a suitable topology for the wireless charging system. The design of the
system principally corresponds to the parameters necessary for resonance operation
are the inductors and capacitors: f 0 = 2π √1C L = 45.96 kHz. The values of the
s s
primary capacitors selected for the two topologies at Cs = 0.31 uf are: for example:
at M = M = 1.97 × 10−5 , d = 30 cm: C p (uF) (Serie Serie) = 0.31 and C p
(uF) (Serie-parallel) = 0.41. Of the four types of compensation, both Serie-Serie
and Serie-Parallel are suitable for building a wireless charging system: Serie-Serie
topology:—Their normalized primary capacity C pn = 1. It does not depend on
Magnetic Chargers in Electrical Models … 1571
the system settings—primary capacity C p does not depend on the load R L . Serie-
parallel topology:—The SP topology may be preferred for resonance as its normal-
ized capacity for resonance C pn increases with K and does not depend on the load
R L .—primary capacity C p depends Just on the air gap d of the system and is inde-
pendent of the load R L . The two parallel-compensated primaries: parallel-Serie and
parallel-parallel are dependent on the load and their normalized primary capacitance
decreases as a function of the magnetic coupling K. So the two topologies serie-serie
and serie-parallel are suitable for the construction of a resonant magnetic coupler
for the recharging of electric vehicle batteries since their normalized capacities are
not depending on k and the load R L . According to Figs. 6 and 7 the performance of
each topology depends on the frequency of the system: −ηss = 60% at f 0 = 45 kHz
−ηsp = 81% at f 0 = 45 kHz. The performance of the parallel-serie topology is
greater than that found in the serie-serie topology when the system is operated at
the resonance frequency. The Serie-parallel topology is convenient for the charging
system of electric vehicles as it gives us an acceptable efficiency about of 81%. Since
the Serie-Serie topology gives us just a value of 60% at the secondary resonance
frequency. The optimization of parameters such as primary capacity and operating
frequency at constant air gap is necessary to provide the maximum power to the
utilization part and to satisfy the energy requirement, solve the issues of batteries
and charging of electric vehicles.
6 Conclusion
In this paper, a new wireless recharging system based on a resonant electrical circuit
of the inductive coupler power transfer technology is studied. The mutual induc-
tance between primary and secondary coils has been calculated and plotted by using
MATLAB for different air gap values between the two circuits. The performance of
a traditional charging system as a function of the air gap is plotted and discussed.
In addition, a wireless recharging system based on a resonant magnetic coupler has
been developed by adding capacitors to compensate the reactive power consumed by
the primary and secondary coils. The performance of a resonant wireless recharging
system as a function of frequency for both types (serie-serie: serie-parallel) is traced
and discussed. The value of the resonance frequency, the values of the added capac-
itors are calculated and optimized to achieve maximum transfer efficiency. Finally,
the system presented can be a good solution for recharging EV batteries and can be
installed in parking lots, motorways to increase the use of EVs.
1572 A. Elkhatiri et al.
References
1. Elwalaty M, Jemli M (2019) A two-rectangular coils wireless charging electric vehicle with
photovoltaic generato. In: International conference on sciences and techniques of automatic
control and computer engineering (STA) (2019)
2. Elwalaty M, Jemli M, Ben Azza H (2020) Modeling, analysis, and implementation of series-
series compensated inductive coupled power transfer (ICPT) system for an electric vehicle. J
Electr Comput Eng 2020:1–10
3. Ahmad A, Alam MS (2019) Magnetic analysis of copper coil power pad with ferrite core for
wireless charging application. Trans Electr Electron Mater
4. Yang G, Liang H (2018) Adaptive frequency measurement in magnetic resonance coupling
based WPT system. J Int Measur Confeder (IMEKO)
5. Hasanzadeh S, Vaez-Zadeh S (2012) Performance analysis of contactless electrical power
transfer for Maglev. J Magnet
6. Kim Ean K, Takehiro I (2013) Impedance matching and power division using impedance
inverter for wireless power transfer via magnetic resonant coupling. IEEE Trans Ind Appl
50(3)
7. Shuo W, David D (2013) Review of wireless charging coupler for electric vehicles. In: 39th
annual conference of the IEEE industrial electronics society. https://doi.org/10.1109/IECON.
2013.6700342
8. Jaegue S, Seungyong S, Yangsu K (2013) Design and implementation of shaped magnetic
resonance based wireless power transfer system for roadway-powered moving electric vehicles.
IEEE Trans Ind Appl 50(3)
9. Wei Z, Chunting Chris M (2015) Loosely coupled transformer structure and interoperability
study for EV wireless charging systems. IEEE Trans Power Electron
10. Soham Ch, Archana I (2017) Design optimization for an efficient wireless power transfer
system for electric vehicules. In: 1st international conference on power engineering
11. La recharge des voitures électriques. https://www.automobile-propre.com/dossiers/recharge-
voitures-electriques. Accessed 09 June 2020
12. Kalwar KA, Aamir M, Mekhilef (2015) Inductively coupled power transfer (ICPT) for electric
vehicle charging. Renew Sustain Energy Rev 47:462–475
13. Wang C-S, Covic GA, Stielau OH (2001) InGeneral stability criterions for zero phase angle
controlled loosely coupled inductive power transfer systems. In: 27th annual conference of the
IEEE industrial electronics society, vol 2, pp. 1049–1054, USA
14. Wang C-S, Stielau OH, Covic GA (2005) Design considerations for a contactless electric
vehicle battery charge. IEEE Trans Ind Electron 52(5):1308–1314 (2005)
15. Hasanzadeh S, Vaez-Zadeh S (2013) Efficiency analysis of contactless electrical power
transmission systems. Energy Convers Manage 65:487–496
Improve the Energy Harvesting
Alternatives Using the Bond Graph
Approach for Powering Critical
Autonomous Devices
1 Introduction
Nowadays, with increasing concern about clean energy demand the vibration energy
harvesting has become great challenges as a harvestable power for the automotive
industry to power smart sensing actuators and self-powered applications. Generally,
Piezoelectric Energy Harvesting (PEH) from ambient vibration energy is nothing but
a scavenger of anybody’s motion from its mean position [1]. Mechanical vibration
can be harvested from different construction such as a vehicle suspension, vehicle
tires, and any other solid system when subjected to the motion on different road
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1573
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_143
1574 S. Touairi and M. Mabrouki
at both of the piezoelectric materials ends [17]. This approach deals to mount the
piezoelectric beams on a vehicle seat suspension system parallel to the driver all
seat system which maximizes the generates voltage difference distribution in the
piezoelectric material. As a result, this strain generated by the suspension defor-
mation with the road’s higher roughness ensures a wide harvestable electric power.
However, simplified system modelling was presented and improved in this section.
The mathematical analysis is considered with one degree of freedom two degrees
of freedom and a large number of degrees of freedom. Figure 1 shows the vehicle
dynamic architecture divided into sub-modules in the proposed complete BG model.
In this study, the vibration of the vehicle movement supposed vertical as well as we
neglect the pitch and the roll movements. The piezoelectric beam (ceramic PVDF)
integrated into the driver’s seat is under disturbance whit complex mechanical proper-
ties. As in previous works (e.g. [17, 18]), the state equations of vehicle displacement
have been calculated as Eqs. (1), (2), (3) and (4) shown. The harvester dynamics
action and piezoelectric hysteresis are supposed to be an absolute generator force
as a vertical motion of the system that gives the assimilation motion for the suspen-
sion of the complete siege. The conductor’s body is modelled with mass, spring and
damper in the proposed BG scheme that exhibits a phase transition caused by the
mass of the vehicle chassis. The calculated equations for the vehicle dynamic during
motion are as follows:
Therefore, the seat suspension model is supposed composed of the driver body
model (Suspended mass ms ) directly attached to the vehicle chassis, the rolling
weight and tire assembly (Unsprang mass mu ). In 20-Sim software input, the sieges
body mass is presented by mf ; mb is the conductor body weight; zu , zs , zf , zc and
zb are the movements of the equivalent masses; the input excitation (road profile)
was noted zr , cs and ks are the suspensions damping and stiffness respectively. Also,
kt and ct are the positions of the compressibility and restraining of the pneumatic
tires, respectively; css , kss and kc are damping coefficients, stiffness parameter of
the driver’ cushion siege suspension, respectively. The factor u signifies the matrix
of the command force applied on the siege suspension. According to this proposed
model, the connection between the piezoelectric actuators PEA actuating strength
and the consequential beam hysteresis displacement becomes linear. This model
increases the harvested power and decrease the use of the battery (long-life cycle)
[13]. However, the standard state equations of the piezoelectric harvester system
were modified by a functional analysis proposed in [14]. The conversion parameters
(of the piezoelectric beam displacement) were estimated in the stiffness matrix of
electrical energy output based on a BG distributed methodology.
This latter facilitates the task by a simple causality inversion and easily solve
the state equations. Regarding the input excitation frequency scale, the piezoelectric
beam operates only with the initial resonant frequency for conductor safety and rides
comfort.
In this section, an explicit formula of the piezoelectric beam modelling has been
proposed. The relationship between medium pressure whit maximum diversion of
homogeneous yarn-beam is analysed at the exert point load in x = 2 * Lpzt . In addi-
tion, a new graphical representation tool is used to model the interaction between flow
connections and extern efforts. In addition, the associated characteristics are calcu-
lated with the motion equations in the equilibrium conditions, through a new graph-
ical representation tool to model the interaction between flow connections and extern
efforts. This flexible model (with multiple degrees of freedom) achieves control
performing adequately for a broad spectrum of road perturbation. This model gives
an agreeable electric energy amount and decreases the road’s disturbance gain (input
Improve the Energy Harvesting Alternatives … 1577
signal w) for the displacement signal (output z) as seen in Eq. (1). The piezoelectric
harvester system is modelized in Fig. 2 as a function of the suspension parameters
proposed in Ref [13]; that estimate the disturbance amount to regenerate large electric
energy.
The harvested power by this model was Eelectric = 48 W during an hour for vehicle
speed equal to VSpeed = 40 km/h. The fixed connection will duplicate the open-circuit
voltage accumulated to a unique factor, and the practical capacitance will be C =
0.5 of the theoretical capacity impedance. The selection BG model utilizes the
collateral joining in order to raise the electrical power quantity. Owing to the fact
that, the analogy connection will be selected for doubling the flow com-pared to a
unique element outline and multiply the efficient electric power (Fig. 3).
The proposed model dimensions require a small beam length, a large thickness
and a minor size to harvest a wide range of vibration energy. The parameters used in
the simulation are presented as follows [13]:
Mvehicle z usn = −Fsv + FP zt X + FP ztY + ctsy (−z 1 + θ1 ∗ v − z ns
f)
(7)
+ctcy (−z 1 − θ1 ∗ w − z ncY
)
MncY z ncY = FP ztY + cwsY (−z 1 + θ1 ∗ v − z ncY
) + kvsY (z cY − z ncY ) (8)
MncX z ncX = FP zt X + cws X (−z 1 + θ1 ∗ w − z ncX
) + kt y (z x y − z nsY ) (9)
1578 S. Touairi and M. Mabrouki
Where, Mvehicle noted the car’s mass (sprung and unsprung), and Kvy is the suspen-
sion stiffness, Tire suspension stiffness is KtX , Sprung damping coefficient CwX and
Tire damping coefficient CvY .
Esj , Qsj , and Gci for each beams PZTj are compact and convex sets with maximum
and minimum constraints give through the functional analysis. In addition, each
IS Pj considered and solved respectively as a transfer function in the closed solutions
intervals. In order to maximize their power production, each piezoelectric beams IS Pj
produces a power amount Esj , QoS qsj , the sum gives the total power Pcj . Formally,
the piezoelectric hysteresis resolution expressed as: H = [M, {Ksj , Qsj , Dcj },{Vi (.)}]
[19].
The solutions are deduced by the following Bond Graph Inverse resolutions
as shown in Fig. 4. This approach gives fast parameters estimation to check the
implementation of the modelling and reduce the mass matrix size. This solution is
promising to enhance the electric power when the vehicle’s mass is bigger.
In this part, the global road disturbances at 15 m/s have been applied to the piezoelec-
tric harvester by COMSOL and MATLAB/SIMULINK software as shown in Fig. 5.
In addition, the vehicle speed according to the piezoelectric harvester deformations
is analysed. The simulation results were obtained for the PZT-7k harvester substrate.
The thermal analysis in Fig. 5 shows a good cyclical waveform and lower hysteresis
level (total piezoelectric patch creeps’ distortion) about 15.5% for PZT-7k compared
with 25.44% for the classical piezoelectric harvester model. Accordingly, the stan-
dard road profile the relative vehicle movement with the highest road amplitudes
which are respectively 0.32, 0.52, and 0.99 m. The harvested power is 306, 430, and
1580 S. Touairi and M. Mabrouki
Fig. 5 A piezoelectric harvester thermal analysis and Electric vehicle input driving
470 mW for a vehicle’s suspended mass of 340 kg and unsuspended mass 180 kg
animated by the random road function for the transversal excitation during an hour
of displacement as seen in Fig. 8. This simulation exposes efficient energy harvesting
for vehicle seat suspension. In fact, when the amount of vehicle speed increases, then
the piezoelectric harvester dimensions decrease and the power bandwidth decrease
(Figs. 6 and 7).
Figures 8 and 9 show respectively: The large road roughness causing wide vertical
displacements, thus increasing the harvester system performances and the harvestable
electric power quantity. Comparing Figs. 8 and 9, it can be clearly noticed that the
proposed approach has great responses compared to the classical model for which the
harvested power and piezoelectric patch have less efficiency. The latter decreases the
vehicle fuel consumption in higher road disturbances. The piezoelectric beam had
suitable dimensions for manufacture and realization thus defined the need for energy
to consume by the functional analysis approach. Those results prove the model’s
uniqueness, following [14], the two homogeneous drivers’ body mass was used to
maximize the producer power and define the optimal weighted sum of harvester
utility functions. As a result, the Y P Z T decreases the vehicle speed H P Z T to further
conducting greater driver comfort for any road surfaces. Figure 5 summarizes the
characteristics of PZT-7K previously analyzed. Finnaly, note that Hpzt increases with
respect to the number of beams PZT and ISPs . The vehicle speed and the piezoelectric
deformation are presented for the first time according to the hysteresis phenomena
as exposed in the electric power nonlinear function Eqs. (5–9).
4 Conclusion
The principal goal of this article is to improve a piezoelectric system built into the
car seat to quantify and convert mechanical vibrations energy of complex equip-
ment (critical devices) and road disturbance. The system gives sustainable electric
power and improves the voltage magnitude with the purpose of increasing the piezo-
electric substances’ abilities. Moreover, this model increased the time life of the
vehicle’s Micro and Nano devices. As shown in the interpretation part the choice
of the piezoelectric harvester’s bond graphs model developed in this paper is very
important to improving the harvested energy. The time spent to reach 302 mW was
20 s on different road test references (50, 70 and 100 km/h) exploiting the beams’
sustainability and simplicity of representation.
As a conclusion, the simulation results confirm that the use of the piezoelectric
harvester attached to the siege suspension ensures excellent electrical power produc-
tion for the autonomous device alimentation. In addition, modelling this system by
the Bond Graph model tool improves the overview of vehicle dynamics. Finally,
the use of the numerical results allowed harvesting vibration energy beneficial for
the seat suspension. Furthermore, this paper gives a simple model that includes a
detailed analysis of piezoelectric hysteresis behaviours and creeps in multiple road
conditions.
References
6. Ruttanatri P, Cole MOT, Pongvuthithum R (2020) Structural vibration control using delayed
state feedback via LMI approach: with application to chatter stability problems. Int J Dyn
Control. doi: https://doi.org/10.1007/s40435-020-00645-8
7. Shao K, Xu Z, Wang T (2020) Robust finite-time sliding mode synchronization of fractional-
order hyper-chaotic systems based on adaptive neural network and disturbances observer. Int
J Dyn Control. https://doi.org/10.1007/s40435-020-00657-4
8. Wei C, Jing X (2020) A comprehensive review on vibration energy harvesting: modelling and
realization. Renew Sustain Energy Rev 74:1–18. https://doi.org/10.1016/j.rser.2017.01.073
9. Xie X, Wang Q (2015) A mathematical model for piezoelectric ring energy harvesting tech-
nology from vehicle tires. Int J Eng Sci 94:113–127. doi: https://doi.org/10.1016/j.ijengsci.
2015.05.004
10. Yuvapriya T, Lakshmi P (2020) Numerical analysis and performance enhancement of active
suspension system using bat optimization. Int J Dyn Control. https://doi.org/10.1007/s40435-
020-00664-5
11. Zerimeche H, Houmor T, Berkane A : Combination synchronization of different dimensions
fractional-order non-autonomous chaotic systems using scaling matrix. Int J Dyn Control.
https://doi.org/10.1007/s40435-020-00660-9
12. Badfar E, Ardestani MA (2020) Design of adaptive fuzzy gain scheduling fast terminal sliding
mode to control the radius of bubble in the blood vessel with application in cardiology. Int J
Dyn Control. https://doi.org/10.1007/s40435-020-00641-y
13. Touairi S, Mabrouki M (2020) Mechatronic modeling and control of energy recovery in
motorcycle tires. In: 2020 IEEE 6th international conference on optimization and applications
(ICOA), pp 1–5
14. Gambhire SJ, Kishore DR, Londhe PS, Pawar SN (2020) Review of sliding mode based control
techniques for control system applications. Int J Dyn Control. https://doi.org/10.1007/s40435-
020-00638-7
15. Ghelichi M, Goltabar AM, Tavakoli HR, Karamodin A (2020) The cost-effective and multi-
objective optimal positioning of active actuators based on tug of war method in seismically
excited benchmark highway bridge. Int J Dyn Control. doi: https://doi.org/10.1007/s40435-
020-00655-6
16. Gopala Rao LVV, Narayanan S (2020) Optimal response of half car vehicle model with sky-
hook damper based on LQR control. Int J Dyn Control 8:488–496. https://doi.org/10.1007/s40
435-019-00588-9
17. Hasan S, Dhingra AK (2020) 8 Degrees of freedom human lower extremity kinematic and
dynamic model development and control for exoskeleton robot based physical therapy. Int J
Dyn Control 8:867–886. https://doi.org/10.1007/s40435-020-00620-3
18. Iggidr A, Oumoun M (2020) Simultaneous stabilization of single-input nonlinear systems with
bounded controls. Int J Dyn Control. https://doi.org/10.1007/s40435-020-00658-3
19. Touairi S, Mabrouki M (2020) Optimization of Harvester System in embedded vehicle
systems via Bond Graph modeling algorithm. In: 2020 IEEE 6th international conference
on optimization and applications (ICOA). pp 1–6
Online Energy Management Strategy
of FC/UC Hybrid Electric Vehicle Based
on Neural Network
Abstract Fuel cell (FC) vehicles seem to be the solution for clean and efficient
transportation. However, the coupling of the fuel cell with another power source
remains essential to overcome its low power density and preserve its lifetime as long
as possible. Thanks to its fast response and high power density, the ultra-capacitor
(UC) can meet a fast and short-duration engine power demand. This paper presents a
power management strategy between the FC and the UC, based on a neural network
(NN). The proposed NN model is trained with wavelet decomposition outputs. To
ensure that both sources operate optimally, the fuel cell power limits as well as the
state of charge (SOC) of the UC are considered.
1 Introduction
Fuel cells have been attracting a lot of attention from researchers and major car
manufacturers thanks to their zero emission of greenhouse gases. A major advan-
tage of fuel cells is that they are small and lightweight. In addition to that, they can
generate electrical energy as long as the fuel is supplied [1]. However, two impor-
tant disadvantages may limit the mass commercialization of fuel cell vehicles. First,
the high cost, which is estimated to be 5 times higher than conventional vehicles,
and second, its durability, which does not exceed 5,000 h, which is equivalent to
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1585
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_144
1586 N. Rifai et al.
150,000 miles in optimal operation [1]. Hybridization of fuel cell with peak power
sources has proven good results to solve these problems. Batteries and Ultra capac-
itors are the most suitable sources for this type of application. An energy manage-
ment strategy (EMS) is therefore necessary to ensure the optimal distribution of
energy between the main source which is the FC and the secondary source which
is the UC. EMSs as reported in the literature can be classified into 4 classes: Rule
based methods, Controller methods, a time-frequency filter methods and optimiza-
tion methods [2]. For rule based methods, we can find EMSs based on fuzzy logic
[3, 4], or Neural network [5] or a combination of both of them such as in ref [3].
For controllers methods, different methods are investigated such as PI controllers,
differential flatness control [7]. The time frequency filter method is achieved using
low pass filters [8], wavelet transform [9, 10], or Fast Fourier transform [11]. Opti-
mization methods are efficient to solve complex management goals such as reducing
cost, minimizing fuel consumption, increasing efficiency or energy sources lifespan.
The problem of these methods is their complexity and computation burden which
is time-consuming. Multiple methods are used for optimization, the most recurrent
ones are the dynamic programming [12, 13] the adaptive optimal control method
[14]. Pontryagin’s Minimum principle approach [15, 16].
In this paper, the fuel cell was hybridized with a UC, since this architecture seems
to have better performances. In fact, the supercapacitor can more effectively support
the FC to meet transient power demand as the UC can be charged or discharged at
a high current, in which the battery is not able to operate. Moreover, highcurrent
charging and discharging of the battery will also reduce its lifetime [17]. The aim
of this paper is to present an energy management strategy based on time-frequency
filter method using Wavelet Transform and a FC/UC limitation check algorithm
which avoid that the fuel cell works out of its comfort zone and control the UC’s
SOC. The outputs of these are used as the database to train a Neural Network. The
purpose of replacing the two algorithms with a neural network is to reduce calculation
time, and also to be able to use the block for online energy control.
The combination of Wavelet Transform and Neural Network has already been used
in literature. However, the energy system architectures and the use of the Wavelet
transform and neural networks are different from their application in this article.
For example, in [18] Authors worked on battery/ultracapacitor energy system, They
used a nonlinear auto regressive NN to predict was used to predict the used the
wavelet transform to predict the vehicle’s power demand, the wavelet transform is
then applied, and the low frequency output is considered as the reference for the
power battery and high frequency output, the reference of UC power. In [19], a
combination of wavelet transform, neural network and fuzzy logic is adopted. Only
the low frequency output of the wavelet transform is used to train the neural network.
The online control is composed of an adaptive linear neural network and fuzzy logic
controller. This EMS is also applied for a UC/battery energy system.
This paper is organized as follows: Sect. 2 presents the energy management
strategy proposed in this paper. Section 3, the simulation of this EMS and the results
are discussed, and finally a conclusion and perspectives are presented in Sect. 4.
Online Energy Management Strategy … 1587
In this study, we propose a real time energy management strategy for a fuel cell
and ultracapacitor energy storage system. The EMS presented in this article is based
on a neural network. The wavelet decomposition combined with a checking limits
algorithm is used offline to train the neural network.
As the fuel cell is unable to respond to fast and abrupt power demands, the main
objective of the EMS would be to avoid rapid demands on the fuel cell but at the
same time to guarantee the traction of the vehicle. As a signal processing method,
wavelet transform is known for its efficiency in dealing with non-stationary signals.
For this, it will be used for sharing power demand between the two energy sources by
decomposing the load power demand to a low frequency signal and a high frequency
signal. The Mexican Hat wavelet function is used. This function is one of the most
used wavelet function for detecting and localizing transients. The complexity of this
function is not critical, since its efficiency has been proved and it used in offline to
prepare the datasets for the neural network.
Continuous wavelet transform (CWT) is defined as below:
+∞
1
Wa,b (x) = s(t). √ .a,b (x).dx (1)
−∞ a
And a is the width of the wavelet function, which will be variable during the
transformation and will be calculated for each unique spectral component. The signal
s(x) is reconstructed from the coefficients W (a, b) using continuous inverse wavelet
transform given by (3).
+∞ +∞
1 x − b dadb
s(x) = W (a, b) √ (3)
0 −∞ a a a2
The detailed description about the proposed wavelet algorithm can be found in
ref [10].
The low-frequency component of the signal is considered as the reference for the
power to be requested from the fuel cell, P f cr e f . The high-frequency component in
1588 N. Rifai et al.
addition to recovered the braking energyare considered for the power as a reference
of the power to be requested from the ultracapacitor, Pscr e f .
Based on the outputs of the Wavelet Decomposition, the limit check algorithm will
allow the fuel cell to operate in its comfort zone which is the linear opertaing region,
and will control the state of charge of the ultracapacitor. Through the respect of
following constraints:
d Pf c
P f cmin < < P f cmax (5)
dt
The power demand of the vehicle is calculated based on four driving cycles. First ones
are the Atremis Driving cycles which are based on a statistical study done in Europe
within the so called Artemis project. ArtRoad cycle that simulates driving on a rural
road and ArtUrban that simulates driving on a rural road were used. The Highway
Fuel Economy Driving Schedule (HWFET) was also used. This cycle represents
highway driving conditions under 100 mph. Finally, The EPA Urban Dynamometer
Driving Schedule (UDDS) which represents city driving conditions. It is used for
light duty vehicle testing. This is a certified driving cycle for verifying emissions
after a cold start.
Online Energy Management Strategy … 1589
Based on the power demand profiles required for these four driving cycles, the
power needed from the FC and the UC are generated by the offline EMS. 70% of
these data are used as datasets for the neural network training. 30% is used for model
validation.
The neural network has three inputs, the first is the load power demand, Second
is a binary coefficient to indicate whether it is in motor or generator mode. Finally,
the SOC of the UC.
As neural networks work best with features that are at a relatively similar scale
and close to the normal distribution, and in order to reduce the training time of the
neural network. The power features (input power, FC and UC power Targets) have
been normalized by subtracting the mean and then scaling to unit variance, which
means dividing all values by the standard deviation:
x − mean
x= (6)
stand deviation
The outputs of the NN will then be rescaled by multiplying with the same standard
deviation and adding the same average used during training.
In this article, a multiple layer neural network was applied. This network is used
as a general function approximator. It can approach any function with a finite number
of discontinuities arbitrarily well, given a sufficient number of neurons in the hidden
layer. Its simple structure makes it a good solution for an online EMS, since it can be
easily implemented on hardware. For this application, The NN has two hidden layers
of 40 and 20 neurons respectively. Hyperbolic tangent sigmoid transfer function
(tansig) is used to allow the network to learn nonlinear relationships between input
and output vectors. These two layers are then followed by an output layer of linear
neurons, purelin transfer function is employed for nonlinear regression.
Weights and bias were updated using BP (back-propagation) which is the most
powerful method to minimize the performance function [20]. In this paper, Trainlm
network training function based on Levenberg-Marquardt optimization is used. It is
the fastest backpropagation algorithm in Matlab toolbox. The NN performance was
evaluated using the Mean square error.
In the next section, the results of the wavelet decomposition coupled with the limit
check algorithm are presented. Then, the results of the training and validation of the
neural network are provided.
The EMS based on wavelet transform coupled with the check limit algorithm was
tested EPA Urban Dynamometer Driving Schedule (UDDS) presented in Fig. 1. The
vehicle on which this study was conducted is a Citroen Berlingo. As shown in Fig. 1.
this vehicle needs an energy system of about 40 kW to propel it. Considering the
maximum power of the fuel cell studied (1.2 kW), we have reduced the scale of the
1590 N. Rifai et al.
Fig. 1 UDDS drive cycle (blue) and power needed for traction of the Citroen Berlingo vehcile
Fig. 2 First plot: Motor Power (red) FC power (black) UC power (Bleu), Second Plot (SOC of
UC)
power profile required to 12.5 of the real scale. The offline EMS results are presented
in Fig. 2. As can be observed, based on the EMS offline, the power demand is covered,
in addition the power delivered by the FC is slow and of low frequency, on the other
hand it does not exceed its limits which in this case are 200 W at the lower bound and
Online Energy Management Strategy … 1591
12 kW at the upper bound. Then, for UC it can be clearly noticed that its dynamic
is much faster, which makes it possible to efficiently respond to a fast transition of
power demand. And finally, the second plot in Fig. 2 represents the SOC of the UC.
This one is as desired between 0.2 and 0.75 which proves the efficiency of the EMS
in controlling the state of charge of the UC.
Based on the results of the four driving cycles ArtUrban, ArtRoad, HWFET and
UDDS. 70% of the data is standardized and used for NN training. Figure 3 presents
the outputs of NN during training. The comparison between the output of the neural
network and the target powers of the FC and UC, which are the outputs of the offline
EMS, led to a root mean square error (RMSE) of 0.040.
30% of the data is used for NN validation. Three inputs are applied. The Power
demanded by the load which is normalized, SOC of the UC and finally alpha the
binary coefficient. At the output of the NN, the powers are rescaled. The result is
shown in Fig. 4. First plot, In blue, the FC power Target output of the Offline EMS
and in red, the power proposed by the NN. In the second plot, in blue the Target power
UC and in black the power proposed by the NN.As can be noticed the feed forward
NN employed is capable of reproducing the work of the wavelet decomposition
algorithms with the verification algorithm. The error between the two (Targets and
NN outputs) is small and does not influence the global rules of the EMS.
4 Conclusion
References
1. Pollet BG, Staffell I, Shang JL (2012) Current status of hybrid, battery and fuel cell electric
vehicles: from electrochemistry to market prospects. Electrochimica Acta 84:235–249
2. Djerioui A et al (2019) Energy management strategy of Supercapacitor/Fuel Cell energy storage
devices for vehicle applications. Int J Hydrogen Energy 44(41):23416–23428
3. Gao D, Jin Z, Lu Q (2008) Energy management strategy based on fuzzy logic for a fuel cell
hybrid bus. J. Power Sources 185(1):311–317
4. Rodatz P, Paganelli G, Sciarretta A, Guzzella L (2005) Optimal power management of an
experimental fuel cell/supercapacitor- powered hybrid vehicle. Control Eng. Pract
5. Zhou D, Al-Durra A, Gao F, Ravey A, Matraji I, Godoy Simões M (2017) Neural network. J
Power Sources 366, 278–291
6. Zhang R, Tao J, Zhou H (2019) Fuzzy optimal energy management for fuel cell and super-
capacitor systems using neural network based driving pattern recognition. IEEE Trans Fuzzy
Syst 27(1):45–57
7. Thounthong P, Pierfederici S, Martin JP, Hinaje M, Davat B (2010) Modeling and control of
fuel cell/supercapacitor hybrid source based on differential flatness control. IEEE Trans Veh
Technol 59(6):2700–2710
8. Oukkacha I, Mamadou BC, Dakyo B (2018) Energy management in electric vehicle based on
frequency sharing approach, using fuel cells, lithium batteries and Supercapacitors. France,
Paris
9. Masih-Tehrani M, Ha’Iri Yazdi MR, Esfahanian V, Dahmardeh M, Nehzati H (2019) Wavelet-
based power management for hybrid energy storage system. J Mod Power Syst Clean Energy
7(4):779–790
10. Rifai N, Sabor J, Alaoui C (2021) Energy management strategy of a fuel-cell electric vehicle
based on wavelet transform. In: Lecture Notes in Networks and Systems, vol 144, Springer, pp
220–235
Online Energy Management Strategy … 1593
11. MacKie DM, Jahnke JP, Benyamin MS, Sumner JJ (2016) Simple, fast, and accurate method-
ology for quantitative analysis using Fourier transform infrared spectroscopy, with bio-hybrid
fuel cell examples. MethodsX 3:128–138
12. Chen Z, Mi CC, Xu J, Gong X, You C (2014) Energy management for a power-split plug-in
hybrid electric vehicle based on dynamic programming and neural networks. IEEE Trans Veh
Technol 63(4):1567–1580
13. Ansarey M, Shariat Panahi M, Ziarati H, Mahjoob M (2014) Optimal energy management in a
dual-storage fuel-cell hybrid vehicle using multi-dimensional dynamic programming. J. Power
Sources 250:359–371
14. Lin WS, Zheng CH (2011) Energy management of a fuel cell/ultracapacitor hybrid power
system using an adaptive optimal-control method. J. Power Sources 196(6):3280–3289
15. Hemi H, Ghouili J, Cheriti A (2014) An optimal control solved by Pontryagin’s minimum
principle approach for a fuel cell/supercapacitor vehicle. In: Proceedings - 2014 Electrical
Power and Energy Conference, EPEC 2014, pp 87–92
16. Ettihir K, Boulon L, Agbossou K (2016) Optimization-based energy management strategy for
a fuel cell/battery hybrid power system. Appl. Energy 163:142–153
17. Thounthong P, Chunkag V, Sethakul P, Davat B, Hinaje M (2009) Comparative study of fuel-cell
vehicle hybridization with battery or supercapacitor storage device. IEEE Trans Veh Technol
58(8):3892–3904
18. Ibrahim M, Jemei S, Wimmer G, Hissel D (2016) Nonlinear autoregressive neural network
in an energy management strategy for battery/ultra-capacitor hybrid electrical vehicles. Electr
Power Syst Res 136:262–269
19. Zhang Q, Wang L, Li G, Liu Y (2020) A real-time energy management control strategy for
battery and supercapacitor hybrid energy storage systems of pure electric vehicles. J. Energy
Storage 31
20. An N, Zhao W, Wang J, Shang D, Zhao E (2013) Using multi-output feedforward neural network
with empirical mode decomposition based signal filtering for electricity demand forecasting.
Energy 49(1):279–288
The Potential Outcomes of Artificial
Intelligence Applied to the Powered
Two-Wheel Vehicle: Analytical Review
1 Introduction
In recent years, the number of motorized two-wheeler accidents has increased signif-
icantly, several causes are singled out for serious motorcycle accidents: Speeding
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1595
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_145
1596 F. Jalti et al.
According to the WHO report in 2017 [9]. Although PTW vehicles are important
for the mobility of goods and people, they also lead to an increased risk of traffic
accidents. The significant number of deaths and injuries associated with this mode
of transport can be explained by several factors, including the proliferation of these
vehicles and the vulnerability of their users. In Fig. 1, by consulting the statistics
relating to fatalities linked to each mode of transport, we can observe, worldwide,
the percentage of PTW users having died in a road accident is the second highest
after cars [9].
With around 516 million PTWs vehicles registered worldwide in 2013, accounting
for 29% of all registered vehicles, the actual number in circulation will only be greater
as a significant proportion of vehicles go unrecorded. Adding to this, according to
WHO’s Global status report on road safety 2015, that 88% of all registered PTWs
vehicles globally in 2013 were in low- and middle-income countries.
The Potential Outcomes of Artificial Intelligence ... 1597
By linking the factors, multiple studies have been conducted to reveal causalities
and correlations in order to identify the most appropriate remedies and actions [10].
Our analysis showed that since the beginning of the 2000s, accident studies have
tended to increase. From the batch of papers, we reviewed, 26 scientific articles
(representing 22%) have dealt with PTW accidentology. Although, the scientific
research has differed on how accidentological aspects have been treated, the common
factor remains human error in the estimation of the danger.
With the increasing popularity of this transport mode and increased casualty levels,
new national attention is currently being given to this area.
Indeed, it is no coincidence that the number of users is increasing (Fig. 2), but it
is rather due to the growth of emerging countries, where we can see the incentive
to this mode of transport given its accessibility with a price of purchase and cost of
ownership quite low, and also the flexibility to overcome traffic jams. Adding to this
that despite the pollution of two motorized wheels being pointed out, PTWs remain
a low-polluting alternative if we associate the track occupancy and consumption
parameters (close to 1l/100 Km for some models) [11].
In terms of safety, the technological advance of the driving assistance systems
initially developed for cars, offers an opportunity for PTWs if we succeed in popu-
larizing these systems at the level of the motorcycle industry. At the same time, the
application of artificial intelligence in this field is very promising.
1598 F. Jalti et al.
Given the weaknesses of the PTW ecosystem, the challenges are faced on several
levels (public policies, social, economic, R&D, etc.). In terms of scientific research,
more than 80% of the batch of papers we reviewed were conducted after 2010,
reflecting the growing interest in the subject.
Before reviewing the state of the art related to the use of AI in PTWs, it’s necessary
to characterize the system made up of the {Driver, Vehicle, Environment}. The rider
continuously participates in the acquisition of information relating to the environment
as well as his role of Navigation and guidance through which he interacts with road
traffic. As for the machine, it mainly participates in the action.
In such a system we can assign three roles (Navigation, guidance and stabilization)
and we can designate 3 activity levels (Perception, Decision, Action). So, one way
to reflect “the level of intelligence” (or the level of automation) of the machine is to
position the system on the chart of Fig. 3.
The work around PTW Vehicles was examined, through a classification according
to field and theme as well as the focus provided by the research concerned.
Concerning the themes, we observe in Fig. 4 that more than 80% of the
articles focus on (in order of importance): ITS (35%), driving behavior (19%),
Accidentology/safety (16%) and finally modeling (15%).
The Potential Outcomes of Artificial Intelligence ... 1599
The field of intelligent transport systems (ITS) brings together all the subjects that
are interested in driving assistance, optimizing the use of transport infrastructures,
improving safety and security as well as the use of artificial intelligence.
Concerning machine learning (ML) techniques, they have been used with success
in a wide variety of fields ranging from computer vision tasks, industry, autonomous
vehicles, etc. [29]. We now recognize two branches in the field of artificial intelligence
which depend on the type of learning, supervised and unsupervised learning [30]. In
supervised learning, models are designed to learn based on the labeled training data.
These trained models are used to make predictions about invisible data. The idea is
to design models capable of learning the patterns, behaviors and characteristics of
the data seen and to apply this knowledge while making decisions in the invisible
data. In unsupervised learning, models are designed to identify relationships and
1600 F. Jalti et al.
60
QTY of PAPERS
50
40
30
20
10
0
DOMAINE
Y = f (X) (1)
Since the shape of the function is usually unknown, the task is to evaluate different
ML algorithms to determine which one is best to approximate an underlying func-
tion. The learning model differs depending on whether it learns from a fixed set of
parameters, regardless of the amount of data fed, so we will call it parametric ML. Or
if it doesn’t make strong assumptions about the shape of the mapping function, in this
case it is a nonparametric ML model. Parametric models are simple, quick to learn,
and work well even if the data is not perfect. However, they tend to be unsuitable and
heavily constrained to a specific shape. On the other hand, nonparametric techniques
can adapt to many functional forms and can lead to more efficient prediction models.
They require a large amount of data to estimate the underlying mapping function,
take longer to practice, and can lead to overfitting.
Besides, the trend on the importance given by scientific research to Artificial
intelligence in the riding assistance for PTWs is at least recent as can be seen in
graph of Fig. 5. In fact, before 2000 we have almost no project interested in the ITS
related to PTWs. But since 2010 we have witnessed the increase in works that have
focused on intelligent systems applied on PTWs. Also, we had to wait until 2015
The Potential Outcomes of Artificial Intelligence ... 1601
to notice a remarkable shift towards AI with 14 articles out of 17 that address the
concept of AI.
In our analysis, on the 57 articles concerning ITS, we classified our batch
according to the affected area in 29 groupings. Where it was found that the percentage
occupied by artificial intelligence represents 28% distributed according to the table
below (Table 1):
The technological advance in terms of intelligent systems that equip PTWs is natu-
rally led by the manufacturing giants. Since the associated intellectual property is
required, we will cite the main areas of development that the sector knows:
4 Discussion
The analysis of the state of the art around the PTW aims to constitute an image of
the current situation of the ecosystem as a whole and in relation to the potential of
AI. The accident side for two-wheeled vehicles was addressed to identify areas of
weakness for this mode of transport. After the picture is completed by the synthesis
of the various intelligent transport systems, the objective is to gauge the level of
technological advance experienced by this transport sector.
Our system being a triplet {Driver, Vehicle, Environment}, the driver influences
the dynamics of the motorcycle by his actions (torque, roll and angle of inclination),
the interaction with the infrastructure is manifested by the tire/road contact. However,
the driver’s interaction with the environment is amplified [2], hence the need to
improve the level of perception of the system by acting on the reliability and the
quality of the information captured (which until today is mainly captured by the
driver).
Nowadays, ITS for PTWs are only accessible for a small percentage of the fleet,
even for an instrument like ABS which must be fitted to practically all cars manu-
factured, the latter is not necessarily accessible for most motorcycles put on the
market.
By examining topics addressing AI, although we can require an investment in
graphics processing, but the use of image processing associated with ML made
it possible to achieve a fairly high performance (latency <= 0.3 s) [14], also for
articles [7, 16, 18] we have successively dealt with: the problem of lateral dynamics,
cognitive load when driving, obstacle detection. These projects were developed to
be embedded and used image processing associated with machine / Deep learning
algorithms,
Also, the machines equipped with AI have been able to achieve very remarkable
performances, indeed if we are able to reproduce the biker robot “Motobot”, which
has learned through machine learning to achieve almost the same performances
of a Human being, we can think of generalizing learning with exchange between
machines. the idea is not to produce a robot for each machine but rather to consider
the machines as sensors and exploit the potential of connection technologies to ensure
complementarity between the different vehicles. In this sense we know an emerging
field, it is that of AI.
This leads us to assert the perspectives that Artificial Intelligence can play in
the development of driver assistance, particularly to improve the management of
problems with loss of control, speed, wearing of helmets, visibility and alertness.
1604 F. Jalti et al.
Given that the software and hardware development framework is nowadays depen-
dent on the manufacturer’s own design, it’s difficult to achieve interesting perfor-
mances and the price will always be exorbitant. But if we had a common frame-
work, we could hope for faster and more interesting development. Adding to this
that connection technologies will allow access to a larger database for an artificial
learning.
Joining the general trend in this field, the use of AI by object recognition remains
the main orientation of the projects developed for the benefit of PTW. Is this the only
lead? definitely no. In fact, the acquisition of data by sensors (whether PTW or other
means of transport) in different forms (image, sound, electrical or electromagnetic
signal,…) can constitute a real-time database.
In our context, the perspectives that we have drawn up will be focused on the
measurement of the roll angle for stability in curves as well as the detection of
obstacles.
References
1. Bekiaris ED (2009) SAFERIDER Project: New safety and comfort in Powered Two Wheelers.
Humans System Interaction, IEEE
2. Kuschefski A (2011) Advanced rider assistance systems for powered two-wheelers (ARAS-
PTW). In: 22nd international technical conference on the enhanced safety of vehicles (ESV),
Germany Paper Number 11–0123, NHTSA
3. Cossalter V (2006) Motorcycle dynamics. Second edition, Copyright by Vittore Cossalter
4. Damon P-M (2017) Lateral & steering dynamics estimation for single track vehicle:
experimental tests. IFAC, pp 2405–8963
5. Slimi H (2012) Système d’assistance a la conduite pour vehicules a deux-roues motorises.
Automatique/Robotique. Universitéd’Evry-Val d’Essonne
6. Manussawee P (2018) Detection of moving vehicle spacing. In: The 18th International
Symposium on Communications and Information Technologies ISCIT, IEEE
7. Wang J (2018) Road targets recognition based on deep learning and micro-doppler features.
In: International Conference on Sensor Networks and Signal Processing (SNSP), IEEE
8. Attal F (2015) Powered two-wheeler riding pattern recognition using a machine-learning frame-
work. IEEE Trans Intell Trans Syst 16(1):475–487. https://doi.org/10.1109/TITS.2014.234
6243
9. Krug E, Billingsley S, Watson B, Irigoyen JL (2017) Powered two- and three-wheelers safety: a
road safety manual for decision-makers and practitioners. 3.0 IGO, ISBN 978–92–4–251192–5,
World Health Organization
10. Sachin K (2018) Responsible factors of powered two wheeler ACCIDENT _ a review. In:
International Conference RICE, IEEE
11. Haworth N (2012) Powered two wheelers in a changing world - Challenges and opportunities.
Accident Analysis and Prevention, Elsevier
12. Nair Pravin R (2011) Performance evaluation of HMM and neural network in motorbike fault
detection system. In: International Conference ICRTIT, IEEE
13. Krishnamoorthy R (2018) Automated traffic monitoring using image vision. In: International
Conference ICICCT, IEEE
The Potential Outcomes of Artificial Intelligence ... 1605
14. Khandelwal Y (2020) Automatic braking system for two wheeler with object detection and
depth perception. In: International Conference ICCES, IEEE
15. Hanhe L (2019) Helmet use detection of tracked motorcycles using CNN-based multi-task
learning. Creative Commons Attribution 4.0 License, IEEE
16. Simon H (2019) Learning a curve guardian for motorcycles. In: Intelligent Transportation
Systems Conference (ITSC), IEEE
17. Ranjana P (2019) Machine learning algorithm in two wheelers fuel prediction. In: International
Conference on Innovations in Information and Communication Technology, IEEE
18. Ryoichi O (2015) Estimation of mental workload during motorcycle operation. In: 6th
International conference on applied human factors and ergonomics AHFE, ELSEVIER
19. Sarikan SS (2017) Automated vehicle classification with image processing and computational
intelligence. In: Complex Adaptive Systems Conference, ELSEVIER
20. Felix Wilhelm S (2020) Detecting motorcycle helmet use with deep learning. ELSEVIER
(2020)
21. Kuo-Wei S (2006) Applying knowledge structure to the usable fault diag-nosis assistance
system: a case study of motorcycle maintenance in Taiwan. Expert Systems with Applications,
ELSEVIER
22. Won M (2019) DeepWiTraffic_low cost wifi-based traffic monitoring system using deep
learning. In: International conference on mobile ad hoc and sensor systems, IEEE
23. Konrad JW (2017) The impact of the modernization of the injection-ignition system on
the parameters of motion of the motorcycle. In: XXI international polish–slovak conference
machine modeling and simulations 2016, ELSEVIER
24. Kuen-Bao S (2005) Design and implementation of a novel hybrid-electric-motorcycle
transmission. ELSEVIER
25. Daniel A-H (2020) Influence of catalyst, exhaust systems and ECU configurations on the
motorcycle pollutant emissions. ELSEVIER
26. Lindsay EN (2019) Do riders who wear an air jacket in equestrian eventing have reduced injury
risk in falls? A retrospective data analysis. J Sci Med Sport ELSEVIER
27. Patrick S (2010) Perspectives for motorcycle stability control systems. ELSEVIER
28. Bobtsov, AA (2013) Control approaches for complicated self-unstable plants with applications
for two-wheel mobile system. In: Conference on Manufacturing Modelling, Management, and
Control, IFAC
29. Jordan MI, Mitchell TM (2015) Machine learning: trends, perspectives, and prospects. Science
349(6245):255–260
30. Mitchell TM, et al (1997) Machine learning. Burr Ridge, IL: McGraw Hill 45(37):870–877
The Design of an EEG Bio-Amplifier
for the Primary Motor Cortex Zone
in 180 nm CMOS Technology
Keywords EEG · AOP · CMOS · CMRR · IRN · The primary motor cortex
1 Introduction
Neuromuscular pathologies affect the nerves associated with the voluntary movement
of muscles. In the case of motor dysfunctions, there is a disruption of the neural
pathways between the primary motor cortex, where movements are initiated, the
spine and muscles. This ensures that the desired voluntary movement is not translated
into actual movement of the limbs [1].
A system for amplifying and extracting signals from the primary motor cortex
is proving to be very interesting for converting the movement intentions into the
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1607
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_146
1608 Y. Laababid et al.
real movement of a prosthesis, for example. To do this, several types of signals can
be used: EEG, ECOG and action potential. Unlike ECOG devices and intercellular
action potentials, EEG systems have the advantage of being non-invasive, as they do
not require surgery and the electrodes are not prone to gliosis, an immune process
that results in scarring on the surface of the electrode over time [2]. In contrast, EEG
systems must acquire signals through the skull and the scalp, thereby limiting spatial
and temporal resolution. This makes signals more difficult to acquire and process to
obtain the desired information, that is, the will to move.
In this sense, this paper proposes an EEG bio amplifier designed for signals from
the area of the primary motor cortex.
The remainder of this work is organized as follows. Section 2 gives an overview of
the proposed EEG acquisition System design as well as the schematics of the building
blocks. Simulation results and comparisons are presented in Sect. 3. Finally, Sect. 4
describes the conclusion and future work.
When acquiring EEG signals, pulses of much greater amplitude than that of the
interest signals can damage the amplifier. These voltage spikes can be caused by
static discharge upon contact between a charged person and the acquisition system
input.
The electrical protection of the patient is also a concern to have. Thus, a protective
circuit is important to prevent damage to the electrical components while ensuring
user safety.
The protection circuit consists of two parallel diodes to limit the amplitude of
the voltage. The diodes chosen are suitable for low amplitude signals (<200 mA).
Since the transition from non-conductive to conductive state is not instantaneous,
signal distortion may be visible from 300 mV [3]. Since the signals of interest are
of amplitude from 25 to 100 μV, they will not be altered by the protection circuit
illustrated in Fig. 1.
When there is no presence of high amplitude unwanted pulses, the diodes do not
conduct and do not affect the EEG signal. On the other hand, when a pulse of more
than 200 mV is present, the diodes start to conduct the current, so that the system
acts as a short circuit in series with a constant voltage source [3].
Resistances are presented at the input to limit the current and thus provide protec-
tion to the patient against the defects of the amplifier. The Association for the
Advancement of Medical Instrumentation (AAMI) has developed a national standard
in the United States on safe current limits for electromedical devices. The leakage
current limit to the patient is 50 μA [4]. If a short circuit occurs between the VDD =
1.8 V power supply of one of the amplifiers and the electrode input, the resistances
The Design of an EEG Bio-Amplifier for the Primary Motor Cortex Zone … 1609
Table 1 Transistors
Transistors Type W/L (μm)
dimensioning
M1, M2 PMOS 1/0.18
M3, M4 NMOS 0.45/1
M5, M9 PMOS 2/0.18
M6 NMOS 0.5/1
M7 PMOS 3/0.18
M8 NMOS 0.45/1
M10 PMOS 1/0.18
U max 1.8 V
R= = = 45 K (1)
If 40 μA
The proposed Operational Amplifier (AOP) was designed with CMOS 180 nm tech-
nology, shown in Fig. 2. It consists of a right bias circuit, a differential stage in the
center and an output stage on the left. To improve the performance of the AOP, a
PMOS cascade current mirror has been added to a standard two-stage configuration.
This arrangement provides high gain, high bandwidth, high slew rate, high stability
and high input impedance [5].
Since noise is a major concern for an EEG signal acquisition system, PMOS
transistors have been chosen as input. This type of transistors generates less 1/f noise
than NMOS [6]. The dimensioning of the AOP transistors is presented in Table 1. The
1610 Y. Laababid et al.
choice of the dimensions of the transistors has been made with particular attention
to the equivalent noise referred to the input to minimize it [7].
2.4 Filters
A right leg subservience system has been integrated into the circuit to reduce
common-mode interference. The human body is a good conductor, it acts as an
antenna which can pick up the surrounding electromagnetic noise and thus interfere
with EEG signals. Since the signals of interest are very weak, this interference greatly
affects the quality of the signal. The subservience circuit of the right leg illustrated in
The Design of an EEG Bio-Amplifier for the Primary Motor Cortex Zone … 1611
the lower part of Fig. 4 makes it possible to cancel this interference and thus obtain
a clean signal at this level. The common-mode voltage of the body is first perceived
and averaged by the resistors R5 and R6. It is then amplified and inverted before
being returned to the body by the E3 electrode. This negative feedback loop reduces
common-mode voltage [3].
The cables connecting the signal sources and the amplification system are often
shielded to reduce the effects of ambient interference. Active shielding consists of
bringing the shielding of cables to the same potential as that of their internal conductor
using a “shield driver” circuit [8]. The latter consists of a follower amplifier allowing
the common mode to be extracted from the input signal and the shielding of the cables
to this potential [9]. This active shielding is represented in burgundy (see Fig. 4).
1612 Y. Laababid et al.
3 Simulation Results
The Protection Circuit has been simulated to verify that high amplitude pulses are
limited in voltage before entering the instrumentation amplifier. The protection circuit
is used to clip the input signal.
Figure 5a illustrates how the number of diodes used for clipping the input signals
of 600 mV and 800 mV is insufficient for this protection operation.
Figure 5b illustrates how when by acting on the number of diodes used, the input
signals of 600 mV and 800 mV are both limited (clipped) to a value of around 450 mV
by the protection circuit.
A protection system makes it possible to limit the amplitude of pulses to protect
the circuit. Additional protection has been added to the system to limit the leakage
current to the patient to 40 μV to comply with safety standards for electromedical
devices.
Fig. 5 a, b. Effect of the protection circuit on the 600 mV and 800 mV input signals
The Design of an EEG Bio-Amplifier for the Primary Motor Cortex Zone … 1613
AC Simulation. Is mainly used to obtain the frequency response of the circuit. This
simulation allows us to determine the differential gain, Phase Margin and CMRR of
the amplifier. The results are visible in Figs. 6, 7 and 8.
The Miller Compensation technique is essential to achieve stability in a closed
loop. For this reason, a compensation capacitor Cc has been added in series with a
resistance Rz to maintain good stability [10].
CL 1
Rz = 1 + × (2)
Cc gm6
Fig. 6 AC simulation
The proposed EEG acquisition system has been tested using simulations with
Cadence software.
AD
CMRR = 20Log (3)
AC
Filter
The filter has also been tested to verify that the bandwidth is between 14 Hz and
30 Hz, the frequency range of the signals of interest in the primary motor cortex (see
Fig. 14 and Table 2).
1616 Y. Laababid et al.
4 Conclusion
The proposed acquisition system could be improved with the use of dynamic
offset cancellation techniques such as automatic zeroing or chopping. This would
reduce the output signal offset with the system proposed in this article. Another
possible improvement would be to reduce the size of the operational amplifiers. To
minimize the noise of the transistors, relatively large dimensions have been chosen,
which means that the overall size of the EEG acquisition system is larger.
References
1. Lena HT, Hillel JC, Randy D (2015) Neuromechanical principles underlying movement
modularity and their implications for rehabilitation. Cell Press 86:38–53
2. Polikov VS, Tresco PA, Reichert WM (2005) Response of brain tissue to chronically implanted
neural electrodes. J Neurosci Methods 148(1):1–18
3. Webster JG (2010) Medical instrumentation. Application and design, 4th edn. Wiley
4. Olson WH (1978) Electrical safety. Med Instrum Appl Des 4:713
5. Gupta P, Mehra R (2016) Low power three-stage operational transconductance amplifier design.
Int J Adv Res Comput Commun Eng 5(7):314–320
6. Allen PE, Holberg DR (2002) CMOS analog circuit design. 2nd edn. Oxford University Press
7. Akhil G, Shweta S (2018) Design of two-stage CMOS Op-Amp with high slew rate and high
gain in 180nm. In: IEEE proceedings of the second international conference on I-SMAC (IoT
in social, mobile, analytics and cloud), pp 341–345
8. Al Nasser E (2014) The stability analysis of biopotential measurement system equipped with
driven-right-leg and shield-driver circuits. IEEE Trans Instrum Meas 63(7):1731–1738
9. Spinelli EM, Reverter F (2010) On the stability of shield-driver circuits. IEEE Trans Instrum
Meas 59(2):458–462
10. Jacob R (2010) CMOS circuit design layout and simulation. 3rd edn. IEEE Press
11. Janaki K, Beno A (2017) Noise immune bio potential amplifier for neural amplifying
applications. Int J Adv Res Electr Electron Instrum Eng 6(5):3310–3316
12. Gubta L, Kumar A (2018) Comparision on low-noise neural signal amplifiers and design of
current mirror OTA for EEG. Int J Pure Appl Math 119(12):14769–14784
13. Chung-Jae L, Jong-In S (2019) A chopper-stabilized amplifier with a tunable bandwidth for
EEG acquisition applications. IEEE Access J 7:73165–73171
14. Akshay G, Gurmohan S (2013) Novel high gain low noise CMOS instrumentation amplifier
for biomedical applications. In: International conference on machine intelligence research and
advancement, vol 83, pp 392–396
15. Manish G, Smriti K (2011) DC suppressed high gain active CMOS instrumentation amplifier
for biomedical application. In: IEEE proceedings of Icetect, pp 747–751
16. Chung-Yu W, Chia-Shiung H (2015) An 8-channel chopper-stabilized analog front-end ampli-
fier for EEG acquisition in 65-nm CMOS. In: IEEE Asian solid-state circuits conference, vol
19, no 5 (2015)
Design of a Non-inverting Buck-Boost
Converter Controlled by Voltage-Mode
PWM in TSMC 180 nm CMOS
Technology
1 Introduction
Renewable energies, such as solar, wind and others, are promising solutions for
ensuring access to clean energy [1]. In several applications related to the production
of electrical power from a renewable source and battery management systems [2], the
DC-DC converters are required to remedy the instability of the produced energy by
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1619
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_147
1620 A. Boutaghlaline et al.
these conversion devices. In the case of solar energy, the photovoltaic cell produces
a DC voltage which varies according to weather conditions and the sunlight incident
angle, to regulate the produced voltage, a Buck-Boost converter is required to satisfy
the need of application [3].
As shown in Fig. 1, a Buck-Boost converter in his conventional topology, uses a
switch and a diode to deliver a DC output voltage with inverted polarity [4], lower or
higher than its input voltage, depending on the switching signal duty cycle controlling
the switch state [5].
To overcome the problem of output voltage inverted polarity where a conventional
Buck-Boost converter is used, another Buck-Boost architecture that combines a Buck
converter and a Boost converter to keep the output voltage polarity same as the input
polarity [5].
As shown in Fig. 2, there are three operation modes where the converter can
operate: Buck, Boost and Buck-Boost. In the case where the input voltage Vi is
greater than the output voltage Vo , the converter operates in Buck mode. When Vi is
lower than Vo , the converter operates in Boost mode. This converter can also work
in Buck-Boost mode when the value of the voltage Vi is the same as the voltage
Vo at the output [3]. A non-inverting Buck-Boost converter can work in continuous
conduction mode (CCM). The current through the inductor flows continuously in
this mode, as shown in Fig. 3 (a), only the duty cycle affects the output voltage.
In discontinuous conduction mode (DCM), the current through the inductor falls to
zero during the off time of the switching signal period, as shown in Fig. 3 (b), so the
output voltage depends on some factors other than the duty cycle, such as the output
current average, the switching frequency and the inductor value [6].
Fig. 2 Buck-Boost operation modes (a) Boost mode (b) Buck-Boost mode (c) Buck mode
Fig. 3 Inductor current (a) inductor current in CCM (b) inductor current in DCM
The limit between these two conduction modes is when the current flowing
through the inductor is canceled just at the time of switching. The chosen value
of the inductor defines the conduction mode. By calculating the critical value of the
inductor where the converter operates within the limit between the two conduction
modes, by choosing a higher inductor value than the calculated value, the converter
operates in CCM, which induces a strong load current. If the value of the inductor
is chosen to be lower than the critical value, the non-inverting Buck-Boost converter
works in DCM, causing a weak load current.
As already mentioned, the duty cycle defines the converter operating mode. There-
fore by setting the appropriate value of the duty cycle, the converter output voltage
reaches the target voltage. In this sense, two control modes exist, voltage-mode
control and current-mode control [7]. Current-mode control takes into account the
output voltage delivered by the non-inverting Buck-Boost converter and the current
flowing through the inductor to control the switches. As for voltage-mode control,
only the converter output voltage is taken into account. Current-mode control has
several advantages in terms of improved stability in closed-loop and rapid tran-
sient response than control in voltage-mode. The advantages of the voltage-mode
control boil down to the simplicity of the feedback loop design and analysis, and a
1622 A. Boutaghlaline et al.
large amplitude of the ramp signal provides noise immunity for a stable modulation
process [6]. Among the voltage-mode control techniques, Pulse Width Modulation
(PWM), which is a control technique that consists of varying the clock pulse width,
which controls the switches. The operation principle of this technique based on a
DC voltage generation representing the error between the voltage at the output of
the Buck-Boost converter and the reference voltage. This error is compared with a
ramp signal to produce a clock signal with a sufficient duty cycle to correct the output
voltage deviation from the target voltage [8]. Another control technique exists, called
Pulse Frequency Modulation (PFM). This method is widely used to improve the effi-
ciency of Buck-Boost converters for low load currents. The PFM control technique
is based on the idea of varying the frequency of the clock signal controlling the
switches, so the power transmitted to the load is linked to the switching frequency
[9]. The PWM control technique is used for the high load current. In the case of
low load current, this technique use affects the efficiency of the converter due to
switching losses, which involves the PFM technique use to increase significantly the
converter efficiency in the low load current case [10].
The objective of this work is to design a non-inverting Buck-Boost converter
controlled by the voltage-mode PWM technique, in TSMC 180 nm CMOS tech-
nology. The input of this converter is derived from a photovoltaic cell. The methods
leading to the realization of this work are presented in Sect. 2. The completed results
are presented and discussed in Sect. 3. Section 4 presents the layout of the designed
Buck-Boost converter. Finally, this work is concluded in Sect. 5.
2 Methods
Figure 4 presents the block diagram of this work. The non-inverting Buck-Boost
converter is designed assuming the CCM. The converter input voltage is derived
from a photovoltaic cell. The clock signal generated by the control circuit, with a
1.8 V amplitude, and a 10 MHz frequency, to control the MOSFET switches. This
study is carried out and simulated with Cadence Virtuoso tool, using TSMC 180 nm
CMOS technology.
As shown in Fig. 5, this converter includes four MOSFET switches. During the ON
time, Q1 and Q2 switches are simultaneously closed, then the inductor starts storing
energy, so Q3 and Q4 switches are open. During the OFF time, Q3 and Q4 switches
are closed, the load capacitor is charged by the energy stored in the inductor, so the
switches Q1 and Q2 are open. The presented converter is designed according to the
following specifications (Table 1):
Design of a Non-inverting Buck-Boost Converter Controlled … 1623
Fig. 5 Non-inverting
Buck-Boost
Ro = 10 (1)
1624 A. Boutaghlaline et al.
Dut y cycle
Vo = × Vi (2)
1 − Dut y C ycle
Vo × Dut y cycle
Cmin = (4)
Ro × f sw × Vo
Pout Vo × i o
η= = (5)
Pin Vi × i i
Voltage-mode PWM control is chosen in this work. Figure 6 shows the PWM
control circuit. The main characteristic of this architecture is the simplicity of its
implementation.
The feedback is provided to the control circuitry from the converter output by
a voltage divider. This scaled voltage is compared with the reference voltage by a
Schmitt trigger based on a two-stage operational amplifier, to generate a DC voltage
qualified as the output voltage deviation from the reference voltage. The obtained
error is compared with a ramp signal using a comparator to produce a clock signal
with an appropriate duty cycle and a switching period same as the ramp signal period
to correct the detected error, as shown in Fig. 7.
Two non-overlapping clock signals are required to control the non-inverting Buck-
Boost converter switches. These two clock signals must not switch at the same time,
to avoid the shoot-through problem over the converter switches. Due to the high input
capacitance of the MOSFET switches, a driver is required to control these switches.
3 Results
This work is implemented and simulated with Cadence Virtuoso tool. The power
converter is designed to deliver an output voltage of 500 mV for input voltages
ranging from 200 to 800 mV.
Figure 8 illustrates the transient response of the designed converter with voltage-
mode PWM control in Buck-Boost mode. The applied input voltage is 500 mV. The
output voltage reaches 501.54 mV and its ripple voltage is 3.53 mV with a 10 load
resistor. The output current average is 50.01 mA. The rise time and the settling time
are 5.46 µs and 6.15 µs respectively. The converter achieves in this mode 83.08%
efficiency.
When the non-inverting Buck-Boost converter operates in Buck mode, with
800 mV as input voltage, the converter provides 526.38 mV with an output voltage
ripple of 4.93 mV, for a resistive load of 10 , as shown. The output current average
is 52.43 mA. The efficiency is 82.90%, with a rise time of 4.66 µs, the settling time is
6.02 µs. The converter output voltage transient response in Buck mode is represented
in Fig. 9.
1626 A. Boutaghlaline et al.
For an input voltage of 200 mV, the designed converter operates in Boost mode.
The output voltage reaches 423.2 mV and its ripple voltage is 2 mV, for a resistive
load of 10 , as shown in Fig. 10. The output current average is 41.92 mA. The
settling time is less than 15.91 µs, and the rise time is 13.69 µs. The converter
efficiency is 70.99%.
The results shows the output voltage delivered by the Buck-Boost converter from
different input voltages (Table 2). In 200 mV and 800 mV as input voltages cases,
the designed converter approached the desired output voltage due to the designed
voltage-mode PWM controller. The controller generates a PWM signal with a closest
duty cycle value to the appropriate one, calculated using Eq. (2). For 200 mV input
voltage, the generated PWM signal duty cycle is 67.91%. For 800 mV input voltage,
the generated PWM signal duty cycle is 39.68%. Using Eq. (2), the appropriate duty
Design of a Non-inverting Buck-Boost Converter Controlled … 1627
cycle for 200 mV is 71.43%. In the case of 800 mV input voltage, the required duty
cycle is 38.46%.
4 Layout
The designed DC-DC non-inverting buck-boost converter and its voltage-mode PWM
control circuitry are implemented in TSMC 180 nm CMOS technology. Figure 11
shows the layout of the design presented in this paper. The layout area of this work
is 0.043 mm2 . The four NMOS switches occupy 65.43% of the total layout area. The
DC-DC power converter layout is done symmetrically, especially for the switching
clocks nets to avoid clock skew. The differential pairs and current mirrors layout is
done respectively with common centroid and inter-digitized matching techniques.
The power nets are routed with a larger width to minimize metal resistivity.
1628 A. Boutaghlaline et al.
5 Conclusion
References
1. Ogudo KA, Umenne P (2019) Design of a PV based power supply with a non inverting buck-
boost converter. In: IEEE PES/IAS PowerAfrica conference: power economics and energy
innovation in Africa, PowerAfrica 2019. Institute of Electrical and Electronics Engineers Inc.,
pp 545–549. https://doi.org/10.1109/PowerAfrica.2019.8928656
2. Ouremchi M, El Alaoui M, Farah F, El Khadiri K, Qjidaa H, Lakhssassi A, Tahiri A (2018)
Li-ion battery charger based on LDO regulator for portable device power management. In:
Proceedings of 2018 6th international renewable and sustainable energy conference, IRSEC
2018. Institute of Electrical and Electronics Engineers Inc. https://doi.org/10.1109/IRSEC.
2018.8702961
3. Ma J, Zhu M, Chen Y, Wang H, Cai X (2019) LPV-based cascade control for three mode
non-inverting buck-boost converter. In: 2019 IEEE 4th international future energy electronics
conference, IFEEC 2019. Institute of Electrical and Electronics Engineers Inc. https://doi.org/
10.1109/IFEEC47410.2019.9015003
4. El Khadiri K, Akhmal H, Qjidaa H (2017) Li-ion battery charging with a buck-boost DC-DC
converter for a portable device power management. J Low Power Electron 13:263–270. https://
doi.org/10.1166/jolpe.2017.1479
5. Bhat S, Nagaraja HN (2015) A novel DSP based non inverting buck-boost converter for PV
system. In: 2015 IEEE international conference on signal processing, informatics, communi-
cation and energy systems, SPICES 2015. Institute of Electrical and Electronics Engineers Inc.
https://doi.org/10.1109/SPICES.2015.7091475
6. Chen WW, Chen JF (2018) Review of the PWM control circuits for power converters. In: Power
Systems. Springer, pp 37–79. https://doi.org/10.1007/978-981-10-7004-4_2
7. El Khadiri K, Qjidaa H (2014) Inverting buck-boost DC-DC converter design challenges. J
Embed Syst Appl 4. https://doi.org/10.5121/ijesa.2014.4101
8. Huang J (2016) A new method of generate the PWM wave using analog circuits. https://doi.
org/10.2991/icsnce-16.2016.23
9. Nagarjun S, Debnath D, Chakraborty C (2018) Buck-boost buck CCM-DCM converter for PV
based DC standalone system. In: Proceedings of 2018 IEEE international conference on power
electronics, drives and energy systems, PEDES 2018. Institute of Electrical and Electronics
Engineers Inc. https://doi.org/10.1109/PEDES.2018.8707723
10. Chen HM, Chang RC, Lin KH (2013) A high-efficiency monolithic DC-DC PFM boost
converter with parallel power MOS technique. VLSI Des 2013:643293. https://doi.org/10.
1155/2013/643293
Low-Power High-Speed On-Chip 5 to 1
Serializer in 180 nm Technology
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1631
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_148
1632 A. Menssouri et al.
EQUALIZER
SERIALIZER FFE DRIVER
CHANNEL
data bus through a deserialization circuit on the receiver block. In this article, we
focus on the serializer [5, 6].
The serializer is used for the parallel to serial conversion of data. There are several
architectures to design it [7]. Figure 2 shows the basic schematic of a 2N:1 serializer
that allows serializing a pair number of data. We may think that a multiplexer can
perform the serialization operation. Still, in practice, we do not get the proper serial-
ization, because different synchronization requirements (setup time, hold time, other
types of delay) impose strict limits on the design. The primary internal blocks are
shown in Fig. 3. Two D flip-flops that transfer input data to the output with a delay
when there is a rising edge of the clock, a D-latch and a multiplexer are required to
design the latter [8].
Our proposed serializer is based on the architecture shown in Fig. 4. It is designed
using D and D-Latch flip-flops but replacing the multiplexer with parallel tri-state
inverters.
D1
2:1 Serializer
D0
Q0
CLK
2:1 Serializer
D3 Q1 CLK P0
2:1 Serializer OUT
D2 2:1 Serializer
Pn CLK
Q n-1
CLK
2:1 Serializer
Dn-1 Qn CLK
Dn 2:1 Serializer
D1
D flip-flop out
2:1
mux
D0
D flip-flop
D-latch
CLK
Usually the multiplexer has pair number bits (2X --> 1) but in our work we need
an odd-number multiplexer, for this we propose to use a parallel tri-state (0, 1, HZ)
inverter, instead of a simple multiplexer in order to realize a serializer (5 --> 1).
Many logic gates require a tri-state output: high, low, and high impedance. The high
impedance state is also called the high Z state and is used to connect many gates
outputs to a single line, such as a data bus or address line. A potential conflict would
exist if more than one door output tried to control the bus line simultaneously. A
controllable high impedance state circuit solves this problem [9]. Figure 5.a shows
the schematic of a tri-state inverter proposed in our architecture.
When the level of the EN input is high (EN = 1), the component behaves like a
typical inverter, its output always gives a level complementary to that applied to the
“IN” input. On the other hand, when the level of the EN input is low (EN = 0), the
“out” output goes to the state called “high impedance”.
Figure 5.b shows the simulation obtained for the three-state inverter such that this
result appears that the output signal “out” (the third waveform from top) is the inverse
of the input signal “IN” (the second waveform from top). When the transmission gate
‘EN = 0’ (the first waveform from top) the output gives a state called high impedance.
A D-Latch primarily generates a delay to the input data. It is based on two tri-state
inverters. When CLK is high, the first tri-state inverter is active and the second is
blocked so that our output Q takes the value of the input D, which is said to be
transparent. If the clock CLK is low the first tri-state changeover switch is blocked
Low-Power High-Speed On-Chip 5 to 1 Serializer in 180 nm Technology 1635
C1
D1
D flip-flop
C0
C2
D2
D flip-flop
C0
C3
out
D3
D flip-flop
C0
C4
D4
D flip-flop
C0
C0
D0
D flip-flop
D-Latch
C0
CLK
and the second one is active then the output Q, in this case, takes the previous state.
Figure 6.a shows the cadence virtuoso implementation of a D-Latch based on two
tri-state [10].
Figure 6.b verifies the logical behavior of D-latch with 0.18 µm technology. We
notice that if the clock “CLK” (the first waveform from top) takes the value “1”
the output “OUT” (the third waveform from top) takes the value of input “D” (the
second waveform from top) and if the clock is low “0” the output is memorizing the
chairman value.
1636 A. Menssouri et al.
VDD
IN OUT
EN
(a)
(b)
D flip-flops are also called “delay flip-flops” or “data flip-flops”. They are used to
store 1-bit binary data. D flip-flops are one of the most commonly used flip-flops as
essential memory and zero-order elements.
Low-Power High-Speed On-Chip 5 to 1 Serializer in 180 nm Technology 1637
CLK
D OUT
CLK
D-Latch
(a)
(b)
Fig. 6 A schematic (a) and simulation (b) of a D-Latch based on two tri-state inverters
Our D flip-flop used are designed by connecting two series D-latches. When the
clock signal is high (rising edge) and if the D input is high, then the output will be
a delayed copy of the input signal, and it will also be high, and if the D input is
low, then the output will become small. If we do not apply any clock input to the D
flip-flop or during the falling edge of the clock signal, there will be no change in the
output. It will keep its previous value at output Q. The output Q, therefore, follows
the input D in the presence of a clock signal [11].
Figure 7.a shows a schematic of the D flip-flop that responds to the clock on its
rising edge.
1638 A. Menssouri et al.
CLK
D flip-flop
(a)
(b)
Fig. 7 A schematic (a) and simulation (b) of a D flip-flop based on two D-latches
Figure 7.b shows the analog simulation waveform of flip-flop D. This analog
simulation shows the output voltage “OUT” (the third waveform from top), the input
voltage “D” (the second waveform from top) and clock voltage “CLK” (the first
waveform from top) as a function of time. If the clock voltage on the rising edge,
the output voltage copies the value of the input voltage otherwise the output voltage
stores the previous value.
Our proposed serializer shown in Fig. 8 was designed under Cadence virtuoso. It
consists of ten D flip-flop, five parallel tri-state inverters and a D-latch. The serial-
ization is done by selection signals generated by a register with ST and RST to rest
Low-Power High-Speed On-Chip 5 to 1 Serializer in 180 nm Technology 1639
circuit, one serial input and parallel circularly shifted outputs. Thus, the first output of
this register is used as the clock signal of the D flip-flop which shift the parallel data
to the parallel tri-state inverters with this high-speed clock signal. In turn, the parallel
tri-state inverters take these shifted data for serialization. The addition of a D-latch
is essential because the data obtained at the output of the D flip-flop is shifted by
small delays which results in the crosstalk of the data. However, the D-latch allows
generating a signal with a large delay to overcome the crosstalk problem. Figure 9
shows simulation result of the serializer block (Tables 1 and 2).
Table 1 Design
Parameter Specification Unit
specifications for proposed
5:1 serializer Supply voltage 1.8 V
Data bus width 5 Bit
Clock frequency 160 MHz
Table 2 Performances
Performance [6] [5] This work
comparison between a
previous and this work CMOS technology 65 nm 180 nm 180 nm
Supply voltage (V) 1.2 1.8 1.8
Type of serializer 4:1 8:1 5:1
Power consumption (mW) 80 1216 2.02
4 Conclusion
References
1. Nivedita J, Radheshyam G (2015) Design of a new serializer and deserializer architecture for
on-chip SerDes transceivers. Sci Res 06(03):81
2. Jerry Y (2015) System-level design, simulation and measurement for high-speed data links, pp
1–3
3. Abdallah A, Mohamed A (2016) High speed serial link design (SERDES) introduction,
architecture, and application. academia.edu
4. Mohammed H, Maher A, Fawnizu A, Israel Y (2012) Design and FPGA implementation of
PLL-based quarter-rate clock and data recovery circuit. In: 4th international conference on
intelligent and advanced systems (ICIAS2012)
5. Vinod K, Shalini C, Monika G (2016) Design of high speed serializer for interchip data
communication with phase frequency detector. Int J Eng Res Appl 6(5):30–40
6. Huang K, Deng L, Ziqiang W, Xuqiang Z, Fule L, Chun Z, Zhihua W (2015) A 190mW 40Gbps
SerDes transmitter and receiver chipset in 65nm CMOS technology. IEEE
7. Tongsung K (2016) High-speed low-power transmitter using tree-type serializer in 28-nm
CMOS
8. Arunthathi G, Umamaheswari K, Vijeyakumar K (2018) Design of CMOS serialiser. J Netw
Commun Emerg Technol (JNCET) 8(6)
9. Jaume S, Charles F (2004) CMOS electronics: how it works, how it fails. IEEE Solid-State
Circuit Society Newsletter, Canada
Low-Power High-Speed On-Chip 5 to 1 Serializer in 180 nm Technology 1641
10. Kiran K, Fazal N (2017) Design of low voltage D-flip flop using MOS current mode logic
(MCML) For high frequency applications with EDA tool. IOSR J VLSI Signal Process (IOSR-
JVSP) 7(4):9–14
11. Casper L, Sorin C, Stamatis V (2004) Single electron encoded latches and flip-flops. IEEE
Trans Nanotechnol 3(2):237–248
Computing and Analyzing Through
Silicon Via (TSV) Noise Coupling
in 3D-IC Structures
Khaoula Ait Belaid , Hassan Belahrach , Hassan Ayad,
and Fatima Ez-zaki
1 Introduction
In the current and next decade, a major trend of microelectronics is to grow the inte-
gration density by utilizing three-dimensional (3-D) integration based on packages.
Presently, 3D structures are achieved by stacking several dies of integrated circuits
(IC), all these dies communicate with other sub-modules via vertical 3-D intercon-
nections. Hence, 3D interconnection is a good solution that helps 3D integration to
achieve the desired performance [1]. Higher power consumption and increased delay
caused by the large resistance and capacitance of the wires are the main problems
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1643
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_149
1644 K. Ait Belaid et al.
that prevent the improvement of the integrated circuits quality [2, 3]. Indeed, 3D-
integration based on TSVs is a good solution for great integration permitting higher
system speed and lower power consumption [4]. 3-D integrated architecture based
on TSVs is an almost new technique and widely used. But these architectures have
some difficulties in their modeling and analysis. Therefore, a precise model of TSV
is crucial to improve 3D design performance. Several papers have proposed models
of TSV interconnections embedded in a substrate. In [5, 6], authors proposed a tech-
nique based on RF characterizations and simulations, which conduct to an analytical
model based on frequency and comprising MOS effect of high aspect ratio of TSV.
Others in [7] gave an equivalent lumped element model for different multi-TSV
structures and introduce closed form equations for resistive, inductive and capacitive
coupling structures. The closed form equations depend of physical dimensions and
material properties. An efficient method was detailed in [8] to model TSV inter-
connections. This procedure occurs an equivalent network parameter which covers
the combined effect of silicon substrate, insulator, and conductors. The method uses
Maxwell’s equation in integral form. Another equivalent circuit model of the chip-
to-chip vertical via, used in high frequency, and based on physical configuration of
the structure was presented in [9, 10]. The parameters of the model were extracted
and analyzed using S-parameters measurement plying a vector network analyzer up
to 20 GHz range.
One of the problems found in the modeling of 3-D technology is the modeling of
the substrate in high frequencies. Because of substrate distribution nature, it cannot
be converted into a dense analytical model. In general, the substrate coupling models
could be extracted from a full 3-D numerical simulation, based on an appropriate
discretization such as Maxwell’ equations [11]. In [12, 13], the box integration could
be used to derive a distributed RC network of the substrate. In this technique, an
equivalent circuit representation of the substrate is extracted using a 3-D rectangular
RC mesh. In [14, 15], to solve the obtained Maxwell’s equations the Boundary
Element Method is applied as an alternate approach.
In relation to the topic, authors in [16–18] proposed a TSV-TSV, and TSV-Active
circuit noise coupling models. Both of these noise coupling paths models were
extracted and analyzed using a 3-D transmission line matrix approach. A model for
coupling noise that takes into account substrate contacts between TSV in bulk CMOS
technologies is proposed in [19]. The suggested model is dense but has reasonable
accuracy for the compact substrate contacts. To study 3-D integrated circuit struc-
tures, based on TSVs, a computation of noise coupling is required [20–23]. Hence,
this work consists to compute this noise in time-domain. First, both models of TSV-
TSV noise coupling and TSV-active circuit noise coupling are presented. Then, a
method to calculate these coupling noises in the time-domain is given. The tech-
nique is the FDTD-1D. The rest of this paper is organized as follows. The TSV-TSV
and TSV-active circuit noise coupling models are presented in Sect. 2. Then, the
computing technique is explained in the same section. The results and simulations
are presented and examined in Sect. 3. The conclusions are drawn in the fourth
section.
Computing and Analyzing Through Silicon Via (TSV) Noise Coupling … 1645
The 3-D ICs contain several TSVs; therefore, TSV-TSV coupling noise is a notable
problem. To compute the TSV-TSV coupling noise, a basic structure was used. This
structure contains two signal TSVs and two ground TSVs, that are linked by a ground
line, as depicted in Fig. 1 [16]. The TSV-TSV coupling noise might be presented
as an RLGC equivalent circuit as explained in [17]. Indeed, the structure in Fig. 1
can be presented by the lumped circuit given in Fig. 2. In this figure, the TSVs
are represented by Ctsv-total, and the substrate among the TSVs is presented by a
capacitance and resistance, except that between two ground TSVs. The ground TSVs
are linked using a metal line.
The lumped circuit model in Fig. 2 could be further reduced to the simplified
equivalent circuit model in Fig. 3. This simplified circuit contains only three essen-
tial elements: substrate capacitance, substrate resistance, and the equivalent TSV
capacitance. This equivalent circuit was studied with several terminations in Sect. 3.
Fig. 2 The equivalent lumped circuit model for coupling noise between two TSVs
In 3D-ICs, different silicon chips; with several MOS; are stacked using TSVs.
Hence, the coupling noise among the active circuit and the TSV must be consid-
ered and modeled. Some difficulties appear, when modeling the TSV-active circuit
noise coupling. An active circuit is composed of many CMOS devices and intercon-
nections. Each CMOS device is divided into substrate contacts well contacts, n-well,
deep n-well and gate metal. In this case, coupling between the substrate contact
and TSV is significant, due to the direct connection via the bulk silicon substrate.
Indeed, a substrate contact can be thought as an active circuit. Therefore, coupling
noise among active circuit and the TSV might be modeled as the coupling among the
contact and the TSV. TSV-active circuit coupling noise structure is depicted in Fig. 4.
In similar manner with the coupling noise between two TSVs, the basic structure
in Fig. 4 can be modeled by a lumped circuit model given by Fig. 5, then, reduced
to the simplified circuit in Fig. 6, which consists of three elements: the substrate
capacitance, the substrate resistance, and the total equivalent TSV capacitance [17].
Computing and Analyzing Through Silicon Via (TSV) Noise Coupling … 1647
Fig. 5 The equivalent lumped circuit model for TSV-Active circuit coupling noise
In general, TSVs are terminated by the Input/output drivers that must be considered.
Consequently, the circuit in Fig. 3 becomes as illustrated in Fig. 7.
The technique used to compute the TSV coupling noise is the FDTD method. By
observing Fig. 7 Eqs. (1)–(9) could be written.
1648 K. Ait Belaid et al.
Fig. 7 The equivalent lumped circuit model for coupling noise between two TSVs with I/O drivers
d V1 (t)
i 3 (t) = C4 (3)
dt
V1 (t)
i 4 (t) = (4)
R3
d V2 (t)
i 1 (t) = C2 (5)
dt
V (t)
i 21 (t) = (6)
R2
Using Eqs. (1)–(9) and replacing each other, then, applying the FDTD-1D method,
Eqs. (10)–(12) are found.
V (n − 1) + DV1 (n)
V (n) = C (12)
− DV1 (n − 1) + E V 1 (n)
Replacing Eqs. (12) and (10) in Eq. (11), the TSV-TSV noise coupling V1 is
obtained as Eq. (13).
− Vin (n) + aV2 (n − 1)
V1 (n) = −H (13)
+ bV1 (n − 1) + cV (n − 1)
Using Eqs. (1)–(9) in addition to (14)–(17) and applying the FDTD-1D method
the noise Vc4 (n) is found as Eq. (18).
D B C
V1 (n) = Vin (n) + Vc1 (n − 1) + Vc4 (n − m)
A A A (17)
G E F
+ Vc4 (n − m − 1) + Vc2 (n − 1) − V (n − 1)
A A A
Fig. 8 The equivalent lumped circuit model for TSV-TSV coupling noise with interconnect line
1650 K. Ait Belaid et al.
In this section, the proposed method (FDTD) is verified using Matlab and Pspice
simulation tools. The objective is to realize the circuits of the TSV coupling on
Pspice and to program the equations found by the FDTD on Matlab. The proposed
FDTD method is based on discretization, indeed a good choice of the discretization
step is necessary. The circuit’s parameters used during the simulations are presented
by Tables 1 and 2.
A trapezoidal signal switching source from 0 to 1.8 V with rising/falling time of
50 ps at frequency 1 GHz was used. The simulated waveforms of each studied circuit
are depicted in Fig. 9, Fig. 10 and Fig. 11. Based on the results, one can see that the
proposed method is in good agreement with Pspice. Hence, the proposed method is
valid, and it can be used for the studied technology. Observing Fig. 9 and Fig. 10
respectively, the peak-to-peak coupling noise in the first case is about 0.15 V. For the
second case, the peak-to-peak coupling noise is about 0.045 V. These results show
that the noise coupling of both paths, the coupling noise between two TSVs and
TSV-Active circuit coupling noise, are significant, must be considered, and reduced.
Observing Fig. 11, the peak-to-peak coupling noise is about 0.15 V. The shapes in
addition to the peak-to-peak coupling noise are similar to those presented in case 1.
The difference is in the disturbances added by the interconnect line. The results show
-0.05
-0.1
0 0.2 0.4 0.6 0.8 1
-9
Time (s) x 10
0.01
-0.01
-0.02
0 0.2 0.4 0.6 0.8 1
-9
Time (s) x 10
0.05
-0.05
-0.1
-0.15
-0.2
0 0.2 0.4 0.6 0.8 1
-9
Time (s) x 10
1652 K. Ait Belaid et al.
that the interconnect line influences the noise coupling by adding some disturbances
as depicted in Fig. 11. The interconnect line can also add reflections that can threaten
the normal functioning of the circuits. The method proposed in this work shows good
results similar to the one proposed and detailed in [20]. This is why it can be adopted
for the modeling of TSV couplings in the time domain.
4 Conclusions
A lot of papers in the literature treat TSV noise in the frequency domain, while,
noise coupling modeling is more important in the time domain. Modeling in the time
domain is more significant and helps to understand the effects of the noise. Hence, for
a good design and construction of 3-D structures based on TSVs, modeling methods
in the time domain must be proposed. In this paper, a technique to model the time-
domain coupling noise in 3D-IC structures is explained in detail. The proposed
method, FDTD-1D, was verified and validated using Matlab and Pspice tools. Then,
a time domain analysis was done for all studied circuits. Based on the found results,
it could be seen that the noise in all studied cases exceeds 40 mV, which means
that TSV-TSV coupling noise and TSV-Active circuit, are significant and must be
reduced, and that the interconnect lines influence the shape of waveforms simulated.
References
9. Ryu C, Chung D, Lee J, Lee K, Oh T, Kim J (2005) High frequency electrical circuit model
of chip-to-chip vertical via interconnection for 3-D chip stacking package. In: 14th topical
meeting on electrical performance of electronic packaging. IEEE, pp 151–154
10. Qu C, Ding R, Zhu Z (2017) High-frequency electrical modeling and characterization
of differential TSVs for 3-D integration applications. IEEE Microw Wirel Compon Lett
27(8):721–723
11. Afzali-Kusha A, Nagata M, Verghese NK, Allstot DJ (2006) Substrate noise coupling in SoC
design: modeling, avoidance, and validation. Proc IEEE 94(12):2109–2138
12. Kerns KJ, Wemple IL, Yang AT (1995) Stable and efficient reduction of substrate model
networks using congruence transforms. In: Proceedings of international conference on
computer aided design. IEEE, pp 207–214
13. Verghese N, Allstot DJ, Masui S (1993) Rapid simulation of substrate coupling effects in
mixed-mode ICs. In: Proceedings of custom integrated circuits conference. IEEE, p 18.3.1–
18.3.4
14. Sedes T (1993) Substrate resistance extraction for physics-based layout verification. In:
Proceeding of workshop circuits systems signal. IEEE/PRORISC, pp 101–106
15. Costa JP, Chou M, Silveria LM (1999) Efficient techniques for accurate modeling and simula-
tion of substrate coupling in mixed-signal IC’s. IEEE Trans Comput-Aided Des Integr Circuits
Syst 18(5):597–607
16. Lee M, Pa JS, Kim J (2014) Electrical design of through silicon via. Springer, Heidelberg
17. Cho J (2011) Modeling and analysis of TSV noise coupling and suppressing using a guard
ring. IEEE Trans Compon Packag Manuf Technol 1:220–233
18. Lim J (2018) Modeling and analysis of TSV noise coupling effects on RFLC-VCO and shielding
structures in 3D-IC. IEEE Trans Electromagn Compat 60:1–9
19. Watanabe M, Karel R, Niioka N, Kobayashi T, Fukase MA, Imai M, Kurokawa A (2014)
Effect of substrate contacts on reducing crosstalk noise between TSVs. In: IEEE Asia Pacific
conference on circuits and systems. IEEE, pp 763–766
20. Ait Belaid K, Belahrach H, Ayad H (2019) Numerical Laplace inversion method for through-
silicon via (TSV) noise coupling in 3D-IC design. Electronics 8(9):1010
21. Xu C, Kourkoulos V, Suaya R, Banerjee K (2011) A fully analytical model for the series
impedance of TSV with consideration of substrate effects and coupling with horizontal
interconnects. IEEE Trans Electron Devices 58:3529–3540
22. Attarzadeh H, Lim SK, Ytterdal T (2016) Design and analysis of a stochastic flash analog-
to-digital converter in 3D-IC technology for integration with ultrasound transducer array.
Microelectron J 48:39–49
23. Beanato G, Gharibdoust K, Cevrero A, De Micheli G, Leblebici Y (2016) Design and analysis
of jitter-aware low-power and high-speed TSV link for 3D-ICs. Microelectron J 48:50–59
Design of a CMOS Bandgap Reference
Voltage Using the OP AMP in 180 nm
Process
Ahmed Rahali , Karim El Khadiri , Zakia Lakhliai, Hassan Qjidaa ,
and Ahmed Tahiri
Abstract This article proposes the implementation and design of a first-order CMOS
Bandgap reference using an operational amplifier with negative feedback to improve
the power supply rejection ratio (PSRR) and reduce the temperature coefficient (TC).
The circuit is designed in 180 nm CMOS process technology and provides a reference
output voltage of 1.2 V over an extended temperature range from −40 °C to 120 °C
with a measured temperature coefficient of 54 ppm/°C. The BGR chip uses a 1.8 V
supply.
1 Introduction
An essential part in the design of most analog and digital electronic systems is
to produce reliable and accurate voltage and current references. A reference in a
circuit provides a stable point which is employed by other subcircuits to achieve
predictable and repeatable results. It is important that this reference point should
be unchanged significantly under different operating conditions. Analog-to-digital
converter, AC-DC or DC-DC converter, and linear regulators are some examples of
circuit applications where voltage reference are intrinsically required.
The Bandgap is an advantageous solution for generating a stable and precise
voltage reference regardless of temperature variations, power supply, and technology
fluctuations. this circuit operates by combining two voltages having a temperature
dependence of opposite signs Fig. 1. The first is the base-emitter voltage VB E of a
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1655
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_150
1656 A. Rahali et al.
thermal voltage at the reference temperature Tr , VG0 − VB E (Tr ) + (η − χ )VTr TTr
is a linear first- order term in temperature. To have a voltage reference, it is necessary
to cancel or reduce the temperature dependence of this term by using VT with a
judiciously chosen multiplicative coefficient.
The voltage Vr e f is written as:
Vr e f = VB E + K VT (2)
VB E0 − VB E1 VT
I1 = = ln N (3)
RO RO
1658 A. Rahali et al.
4 Simulation Results
5 Conclusion
References
1. Jacob Baker R (2010) CMOS Circuit Design, Layout, and Simulation. Third Edition. Wiley-
IEEE Press
2. Razavi B (2001) Design of Analog CMOS Integrated Circuits. McGraw-Hill Higher Education,
international Edition
3. Akshaya R, Yellampali Siva S (2017) Design of an improved Bandgap Reference in 180
nm CMOS process technology. In 2nd IEEE International Conference on Recent Trends in
Electronics Information & Communication Technology (RTEICT), pp 521–524. India
4. Rincon-Mora GA (2002) voltage reference from diodes to precision high-order Bandgap circuits.
Wiley-IEEE Press
5. Zhao Y, Yue S, Bian Q (2008) A novel low temperature coefficient Band-gap reference without
resistors. In International Multiconference of Engineers and Computer Scientists 2008, IMECS,
vol II, Hong Kong
6. Shrivastava A, Kaur A, Sarkar M (2017) A 1.2 V, 33 ppm/°C, 40 nW, regeneration based BGR
circuit for nanowatt CMOS. In International SoC Design Conference (ISOCC), pp 111–112,
South Korea
Mechatronic, Robotic and Control System
Analytical Study of the CMOS Active
Inductor
Imane Halkhams, Wafae El Hamdani, Said Mazer, Moulhime El Bekkali,
and Mohammed Fattah
1 Introduction
Frequency filtering is a key element in electronic circuits, which explains the number
of studies carried out on filtering operations [1, 2]. In the literature, different topolo-
gies of active inductors have been presented. Widely used in modern telecommu-
nication systems such as the mobile phone network, wireless and satellites. Their
manufacture has undergone a vertiginous evolution. Indeed, integrable filters are
produced in tens of millions each year and are present in all daily telecommunication
equipment (smartphones, TVs, ADSL, etc.).
Active inductor-based filters reported in the literature [3, 4], have many advantages
such as frequency tuning ease, high-quality factors, and on-chip integration, however,
the transistor’s parasitic effects impact the linearity and stability of systems which
I. Halkhams (B)
SED, Research, Systems and Sustainable Environments Laboratory, The Private University
of Fez, Fez, Morocco
e-mail: [email protected]
W. El Hamdani · S. Mazer · M. El Bekkali
IASSE Laboratory, Sidi Mohamed Ben Abdellah University, Fez, Morocco
M. Fattah
IMAGE, Moulay Ismail University, Meknes, Morocco
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1665
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_151
1666 I. Halkhams et al.
calls for accurate noise and linearization studies. In this context, we propose a study
of the intrinsic parameters of the CMOS transistor to optimize them and to provide
adequate power to the active inductor circuit. We also propose an analytical study of
the latter to highlight its role in the frequency filter chain.
We propose a study of the active inductor described in [5], to facilitate the design
of bandpass filters and LNAs. This article starts with the extraction of the intrinsic
parameters of the transistor via the h and y matrices. In the third section, we present
the analytical study of the active inductor, using the parameters calculated in Sect. 1.
The need for a high inductance value at the resonance frequency requires that the
transconductances used in the active inductor design be properly configured. To do
this, a small signal study from the transistor is fundamental. The term small signal
is dedicated to signals small enough that the characteristics of the components vary
slightly and can remain in a linear approximation. We used a CMOS transistor whose
five-element equivalent model is given in Fig. 1.
In this model, the components electrically characterize the transistor’s frequency
behaviour. The channel is described by its gds conductance. The gm transconduc-
tance of the transistor determines the gains. The gate-channel interaction is mainly
capacitive because of the existence of the MOS conjunction.
2.1 H Parameters
H parameters are hybrid parameters that connect the input voltage VGS to the input
current IG and the output voltage VDS to the output current ID. The equations
according to the H matrix are written:
Parameter H12 represents the coupling of the output to the input. To calculate it,
this time we inject a voltage at the output and we measure the voltage at the input of
the circuit.
V gs
H12 = =0 (3)
V ds
Parameter H21 is the current gain of the transistor. To calculate it we inject a
voltage at the input while short-circuiting the output voltage, then we calculate the
currents IG and ID.
H 21 = II GD
(4)
gm
H 21 = j.ω.(Cgs+Cgd)
At the transition frequency, the current gain of the transistor H21 is equal to the
unit.
|H21| = 1 (5)
2.2 Y Parameters
To know the functionality of a quadruple, one method consists of studying its voltage-
current transformation Y (matrix admittance), defined by the Eq. (6):
⎧
⎪ Y 11 = |V 2 =0
I1
⎪
⎨ V1
Y 12 = I1
V2
|V 1 =0
(6)
⎪
⎪ Y 21 = I2
|V 2 =0
⎩ V1
Y 22 = I2
V2
|V 1 =0
1668 I. Halkhams et al.
To do this, we proceed by making each time short circuits for the admittance
matrix. However, the realization of short circuits becomes almost impossible when we
exceed 100 MHz, because of the parasitic capacitors and inductors, hence the interest
to use the S matrix, which is very effective for high-frequency implementations and
which is measurable by loading the input and the output with a load of 50 Ω.
By studying the small-signal circuit of the transistor, the latter’s equations using
the Y representations are given by:
I G = Y 11.V gs + Y 12.V ds
(7)
I D = Y 21.V gs + Y 22.V ds
IG
Y 11 = (8)
V gs
IG
Y 12 = (10)
V ds
Y 12 = − j.ω.Cgd (11)
ID
Y 21 = (12)
V ds
Y 21 = gm − j.ω.Cgd (13)
ID
Y 22 = (14)
V ds
Finally, from (9, 11, 13 et 15) we can extract the intrinsic parameters of the
transistor:
Analytical Study of the CMOS Active Inductor 1669
(16)
∂Id(Vgs, Vds)
gm = (17)
∂Vgs
The output conductance gds represents the resistance of the channel. The greater
the width of the area under the gate, the smaller the conductance. The space load
zone is an area in the vicinity of the metal-semiconductor interface. It is less rich in
electrons than in the volume of the semiconductor.
∂ I d(V gs, V ds)
gds = (18)
∂ V ds
The Cgd capacitor is related to the variation of the load stored under the gate
according to the gate-drain voltage when the gate-source voltage is fixed:
∂ Qg(Vgs, Vgd)
Cgd = (19)
∂Vgd
The Cgs capacitor is linked to the variation of the load stored under the gate
according to the source-gate voltage when the gate-drain voltage is fixed.
The Cds capacitor takes into account the electrostatic coupling between the heavily
doped regions below the source and drain contacts.
The transition frequency is a characteristic of the transistor’s operating speed.
Because the higher the frequency, the more the transistor can achieve a high transcon-
ductance for a low capacity. The transition frequency can be calculated from the
current gain of the transistor.
gm
fT ≈ (21)
(Cgs + Cgd)
The calculated parameters will be used in the study of the active inductor in the
following paragraph.
1670 I. Halkhams et al.
The gyrator circuit was invented by Bernard D. H. Tellegen (1900–1990) [6]. Its
symbol is given in Fig. 2. It was first designed for low-frequency applications
where transconductance operational amplifiers were used to simulate active inductors
connected to the mass or floating using external integrated capacitors.
The gyrator considered a 2-port component, when closed to an impedance (Fig. 2),
is described by the following impedance matrix:
V1 I1 Z11 Z 12 I1
= [Z] = (22)
V2 I2 Z 21 Z 22 I 2
Thus:
This equation states that there is no energy generated, stored, or dissipated since
the input power is equal to the output power.
This equation means that when the ideal gyrator is closed on a capacitor, it
simulates the effect of an inductive impedance of value:
L = Rg2 .C (27)
The active inductor consists of a common source transistor, in feedback with another
transistor mounted in common gate. The two transistors form a gyrator that trans-
forms the internal Cgs1 capacitor (between gate and source) of the first transistor
into an inductor.
The small equivalent signal model of the inductor used in [7] is given in Fig. 3.
The Yin input admittance calculated from Fig. 3(a) is given in (28):
gm1.gm2
+ gm2.gds1 + gds1.gds2 + R −1 (gm1
+ gds1 +
gds2)
+p. Cgs2(gm1 + gds1 + gds2) + Cgs1 gds1 + R −1
+p2(Cgs1.Cgs2)
Yin = (28)
gds1 + R −1 + p.Cgs2
Using the small-signal model of the active inductor and taking into account the
parasitic capacities, input admittance calculated from Fig. 3(b) is given in (29).
(a)
(b)
Fig. 3 Active inductor small signal equivalent model (a) simplified (b) with capacitors
1
Yin = gm1 + j.w.Cgs1 + gds1+R −1 Cgs2
(30)
gm1.gm2
+ j.w. gm1.gm2
1 1
Yin = + Cp + (31)
R Rs + L
Analytical Study of the CMOS Active Inductor 1673
By investigating Eq. (30) and by identifying between (30) and (31), four terms
can be deduced (32) [6]:
⎧ Cgs2
⎪
⎪ L = gm1.gm2
⎪
⎨ −1
Rs = gds1+R
gm1.gm2 (32)
⎪
⎪ Rp = gm1
1
⎪
⎩
C p = Cgs1
This proves that the active inductor is equivalent to a resonator, which can be used
for band-pass filtering applications as well as for oscillator design.
The resonator circuit may be considered as an inductor if the operating frequency
is less than the resonance frequency. The two parallel and serial resistors are the
sources of resonator circuit losses. To remedy this, a resistor loss compensation
circuit shall be added to the active inductor circuit.
4 Conclusion
A CMOS active inductor analytical study was presented in this paper to understand
its functioning. H and Y parameters were used to extract the transistor’s intrinsic
elements.
Active transistor-based inductances only and using internal capacity have been
a giant step towards systems integration and in the implementation of preselection
filters, bandpass filters, LNA, and oscillators. All these implementations exploit the
inductive effect and the selectivity around the resonance frequency.
This analytical study demonstrates the functioning of active inductors using
CMOS transistors. The simplicity of their configuration makes them indispensable
in the implementation of bandpass filters that exploit resonance to select a frequency
band [8]. However, the selection of a restricted frequency band or even a single
frequency requires a very narrow bandwidth, characterized by a very high-quality
factor, which leads to selectivity and loss study to make bandwidth as narrow as
possible. This study may be used in the design of bandpass filters and LNAs [9–11].
References
1. Erokhin VV et al (2019) Active tunable bandpass filter with voltage gain control in SiGe
BiCMOS 130 nm. In International Siberian Conference on Control and Communications
(SIBCON)
2. Pantoli L, et al (2017) Design considerations and effects of class-AB polarization in active filters
realized by means of active inductors. In: 47th European Microwave Conference (EuMC), pp
37–40
1674 I. Halkhams et al.
3. Wu Y, Ismail M (2003) RF bandpass filter design based on CMOS active inductors. IEEE Trans
Circuits Syst 50(12):942–949
4. Leoni A, et al (2017) Bandpass filter design with active inductor by means of wave digital
approach. In 13th International Conference on Advanced Technologies, Systems and Services
in Telecommunications (TELSIKS), pp 339–342
5. Halkhams I et al (2017) A Selective active filter for the 5G in the mm-wave band in pHEMT
technology. Contemporary Eng Sci 10(4):193–202
6. Tellegen BDH (2014) The gyrator, a new electric network element. Philips Res. Rep. 3: 81–101.
April 1948, archived from the original on 2014–04–23. Retrieved 20 Mar 2010
7. Halkhams I, et al (2017) Design of a dual-band bandpass filter using active inductor principle.
Int J Commun Antenna Propagation (IRECAP) 7(4) ISSN 2039–5086
8. Ren S, Benedik C (2013) RF CMOS active inductor band pass filter with post fabrication
calibration. AEU - Int J Electr Commun 67(12):1058–1067
9. Nawaz AA, et al (2019) A Ka/V band-switchable LNA with 2.8/3.4 dB noise figure. IEEE
Microw Wirel Compon Lett 29(10)
10. Lee W, Hong S (2020) 28 GHz RF front-end structure using CG LNA as a Switch. IEEE
Microw Wirel Compon Lett 30(1)
11. Mudavath M, et al (2020) Design and analysis of CMOS RF receiver front-end of LNA for
wireless applications. Microprocess Microsyst 75
SM Versus PI Tunning for PWM Rectifier
for SEIG in Wind Energy Application
Abstract This work presents the sliding mode control applied on the self-excited
induction generator, typically used in electrical energy production and particularly
in wind turbines. The machine is associated with an active load through a bank
of capacitors and operated as a generator. Its stator is associated with the AC/ DC
converter. The SM control is presented by a cascade structure that includes tow loops:
The internal loop controls the machine currents “id and iq” and the external loop
controls the Vdc voltage. The whole model (SEIG/PWM rectifier/load) is simulated
using the Matlab/Simulink software, then the results of this simulation validated
the SM superiority over the typical PI regulator. The proposed SM method hold
various advantages: it is recommended to control the systems that contain multy-
variables, for both linear and non-linear systems, it provides great transient and
steady performances, it’s robust against external disturbances, robust against uncer-
tainties/parameters variations, and offers lower THD harmonics versus the typical
PI.
1 Introduction
The wind generation system has gained recently a great importance as a sustainable
and secure source of renewable energy. A three-phase asynchronous machine could
be operated as an asynchronous generator, either connected to the electrical network
or operated in isolated power generation mode with a three-phase capacitance coupled
to the stator windings [1]. This type of generator operated in excitation autonomous
mode, called self-excited induction generator is the most useful generator for remote
windy aeras because it doesn’t require an external power supplies to produce the
excitation magnetic field, the excitation is delivered by supplemental stator terminal
excitation capacitor bank. These references [2–4] have presented several algorithms
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1675
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_152
1676 O. Fadi et al.
and approaches to determine the minimum value of the excitation capacitor. SEIG
has a several advantages such as simple and robust construction, minimal mainte-
nance, absence of DC power supply for excitation and inherent overload protection
[1]. However, the principal drawback of such generators is the voltage control that
can be highly complex depending on the desired performance, this complexity is
more precisely located at the power converter that is unable to produce a cleaned
and continued DC voltage waveform, this issue is mainly caused by the analytical
nonlinear model of the machine that hold numerous variables and that is very respon-
sive to external disturbances (wind speed and load changes) [1]. The control of these
power converters to maintain the output dc voltage regulated to a desired value and
to generate the stator currents with a lowest possible harmonic distortion has been
widely studied by engineers and researchers. The Conventional techniques of PI
control, cover a wide range in industrial applications. These are linear control tech-
niques and have the advantage of simplicity of implementation and ease of synthesis.
Over time, the PI control applications will be inefficient, especially if the processes
to be controlled have complex and non-linear structures [4]. Likewise, the PI tunning
is used in a cascaded configuration [5]. In such a configuration, the internal loop is
often accomplished by current controller, as well as the controller in the external
loop regulates dc-link output voltage. There are typically two classical families of
these control configurations [6, 7]: the VOC (Voltage oriented control) similarly as
in VFOC (Virtual Flux oriented control) of induction generator, they are based on
the coordinate transformation between stationary α-β and synchronous rotating d-q
reference system. Both strategies guarantee rapid transient response and high static
results via the internal current control loop. however, these configurations exhibit
some disadvantages such as [6]: A coupled control between the active and reac-
tive components of the current; A coordinate transformation and PI controllers are
required. Another proposed method based on instantaneous direct active and reac-
tive power control loops, called Direct Power Control (DPC) is developed in [8].
This method constitutes a viable alternative to the typical vector control methods
(VOC-FVOC). However, the DPC has several disadvantages [9]: Variable switching
frequency, High values of the line inductance and sampling frequency are needed to
obtain an acceptable current waveform, also power and voltage estimation should be
avoided at the moment of switching (it yields high errors).
All these techniques mentioned above are employed when the controlled part
of the machine is minimally disturbed, they may be sufficient if the requirements
on the accuracy and performance of the system are not too stringent. However, in
the opposite case and particularly when the controlled part is subjected to strong
nonlinearities and to temporal variables, it is necessary to design control algorithms
ensuring the robustness of the generator with respect to the uncertainties on the
external parameters and their variations [10].
The sliding mode tunning is a robust control method [10] that has indisputable
advantages for the poorly identified system or with variable parameters. This paper
is divided into 4 important sections: in Sect. 1, the modeling of the asynchronous
generator, powered by a tow-level inverter/rectifier system is presented. In Sect. 2, the
basic principle of sliding mode is treated and detailed by mathematical equations. In
SM Versus PI Tunning for PWM Rectifier for SEIG ... 1677
Sect. 3, the internal and external regulation loops used the conventional PI controllers
are replaced by the regulators based on sliding mode, the simulation results will be
exposed to observe the validity of the PI compared with SM regulators. finally, in
Sect. 4, a recap of the main points of this study are discussed.
The system constitution based on a squirrel cage induction generator, prime mover,
excitation capacitance and a PWM rectifier connected to an active load [8].
disq 1
= −Ls Rs isq − wr Lm
2
isd + Lm Rr iqr − wr Lm Lr idr + Lr Vsq (1)
dt Ls Lr − Lm
2
disd 1
= −Lr Rs isd − wr Lm
2
isq + Lm Rr idr + wr Lm Lr iqr + Lr Vsd (2)
dt Ls Lr − Lm
2
dirq 1
= Lm Rs isq + wr L m L s isd − Ls Rs iqr + wr Ls Lr idr + Lm Vsq (3)
dt Ls Lr − Lm
2
dird 1
= Lm Rs isd − wr L m L s isq − Ls Rr idr − wr Ls Lr iqr + Lm Vsd (4)
dt Ls Lr − Lm
2
dVsd 1
= (isd − id ) (5)
dt C
dVsq 1
= isq − iq (6)
dt C
The system dynamic of the three-phase two-level rectifier can be expressed using
the park transformation, in dq synchronous reference frame as [8]:
diq
L = Vqs − Riq − Lwid − Vrq (7)
dt
1678 O. Fadi et al.
did
L = Vds − Rid + Lwiq − Vrd (8)
dt
dVbus Vbus
Cbus = ibus − il = Fd id + Fq iq − (9)
dt RL
Where [X] Rn is the state vector, and [U] Rm is the control vector, with n > m.
Generally, the choice of the sliding surfaces number is equal to the dimension of
the control vector (U) [11]. In order to ensure the convergence of a state variable X
towards its reference value X *, several works propose the following general form:
r−1
d
S(x) = ( + γ) ∗ e(x) (11)
dt
γ : Positive gain; e(X) = X - X*: The error of the variable to be regulated; r:
Relative degree, it is the smallest positive integer representing the number of times
of derivation to obtain the control (d S)/dt=0 ensuring the controllability; S(x): is
an autonomous linear differential equation whose response e (x) tends to zero for a
correct choice of gain γ and this is the main objective of this control.
The condition of the convergence is the criteria which allow the dynamic of the system
converges towards the sliding surface and remains constant against the disturbances
SM Versus PI Tunning for PWM Rectifier for SEIG ... 1679
[10]. For this reason, A candidate Lyapunov function V (x) > 0 is chosen for the system
state variables and a control law is determined to make this function decreases V(x)
< 0. The appropriate Lyapunov function for the system is expressed as follow [10]:
1
V(x) = S(x)S(x)2 (12)
2
By deriving this one, we obtain:
˙ = S(x)S(x)˙ < 0
V(x) (13)
The law control is necessary to attract the trajectory of the desired state towards the
surface and then towards its equilibrium point while ensuring the existence conditions
of the SM [10]. The law control structure has two parts, the first concerns the exact
linearization Ueq and the second Un is a stabilizer part, this one is used to eliminate
the model uncertainties and to dismiss the external disturbances [10].
The two parts expressions of the control U is determined by the set of equations:
dS ∂S ∂x
Ṡ = = . (15)
dt ∂x ∂t
By replacing (10) and (13) in (14), we find:
∂S ∂S
Ṡ = [A][X] + [B]Ueq + [B]Un (16)
∂x ∂x
Replacing the equivalent control by its expression in (15), we obtain the new
expression of the derivative surface [10]:
˙ = ∂S ∗ [B] < 0
S(x) (18)
∂x
1680 O. Fadi et al.
In order to satisfy the condition, the sign of Un must be opposite of S(x) = ∂S∂x
∗[B].
The simplest form that the discrete control can take, is the ‘sign’ function presented
as:
Un = Kx ∗ sing(S(x)) (19)
In this study, the SM technique is applied to the asynchronous generator and the
control expressions are established based on the following layout model:
Figure 1 shows the cascade configuration of sliding mode regulator employed
to the SEIG. The external loop output variable presents the current references of
the internal loops [10]. The references output are the direct and quadrature control
voltages to be applied (Vrd ,Vrq ) to the PWM rectifier. The cascade configuration
imposes a choice of three aeras along the two axes (d,q). The internal loops regulate
the currents “id and iq” and the external loop regulates the Vdc voltage.
eq di ∗d
Vrd = Vds − Ri d + wLi q − L (24)
dt
The discontinuous component of the voltage is given by:
∗
Vrd = −K d ∗ sat(Sd ) (25)
By following the same approach detailed above, the equivalent component of the
voltage Vrq∗ is given by:
di∗q
eq
Vrq = Vqs − Riq − wLid − L (26)
dt
The discontinuous component of the voltage is:
∗
Vrq = −K q ∗ sat(Sq ) (27)
1682 O. Fadi et al.
eq Vbus
ibus = (32)
RL
Vbus ∗
Pbus = Vbus *ibus = Vq *iq ⇒ i∗q = *i (33)
Vq bus
Figure 2 presents the dc bus voltage across the rectifier terminal; the regulation of
this bus is examined by both SM and PI controllers. In spite of the presence of load
step changing and, as illustrate in the figure below, the SM performances of the DC
voltage is better than PI performances, in termers of, the robustness against the load
disturbances, also, the dynamic property is better: there is no overshoot and the SM
response time is smaller than the classical PI time response.
Figure 3 demonstrates the performances of active and reactive powers in the
presence of load variation. During the steady regime, the P power is maintained
constant, and the Q power is maintained at 0 KVA which lead to obtain a good
SM Versus PI Tunning for PWM Rectifier for SEIG ... 1683
power factor (PF = 1), furthermore, The SM ensures a good performance than the
conventional PI regulator (smaller overshoot, smaller response time and constant
steady state).
Figure 4 shows the steady state for the SEIG currents, using both SM and PI control
strategies. As can be observed, the SEIG currents, in SM regulator technique, are
close to sine wave better more than the PI regulator technique. Furthermore, the
Fig. 5 shows that the SM exhibits the best power quality (THD = 1.96%) and, the
PI leads to a dispersed harmonic spectrum with a higher THD of around 6.24%.
6 Conclusion
This work proposes a nonlinear control of the self-excited induction generator using
SM approach. In the suggested control technique, stable and independent controls
of the direct and quadrature currents are achieved. This method can be easily used
in the isolated wind power systems. Furthermore, the simulation results demonstrate
that the dc voltage obtained with the SM tunning is faster compared to the clas-
sical PI. Furthermore, the SM controller brings remarkable improvements compared
to the classic PI regulator since the SM regulators offer good static and dynamic
performances (stability, rapidity and precision), that is to say a shorter response time
without overshoot, also they grant better tracking and rejection of load disturbances,
and reduce THD of SEIG currents.
References
1. Kalaivani C, Rajambal K (2020) Modeling of an efficient high power wind energy conversion
system using self-excited multi-phase machines. Microprocess Microsyst 74:103020. https://
doi.org/10.1016/j.micpro.2020.103020
2. Ouafia F, Ahmed A (2018) Elaboration of the minimum capacitor for an isolated self excited
induction generator driven by a wind turbine. In Conference: International Conference of
Computer Science and Renewable Energies, Ouarzazat
3. Chaturvedi Y, Kumar S, Gupta V (2020) Capacitance requirement for rated current and rated
voltage operation of SEIG using whale optimization algorithm. Procedia Comput Sci 167:2581–
2589. https://doi.org/10.1016/j.procs.2020.03.315
4. Silva EO, Vanço WE, Guimarães GC (2020) Capacitor bank sizing for squirrel cage induction
generators operating in distributed systems. IEEE 8:27507–27515
5. Wang F, Zhang Z, Mei X, Rodríguez J, Kennel R (2018) Advanced control strategies of
induction machine: Field oriented control, direct torque control and model predictive control.
Energies 11(1):120
6. Toufik A, Taibi D, Oualid A (2018) Voltage oriented control of self-excited induction generator
for wind energy system with MPPT. In AIP Conference Proceeding
7. Abdelrahem M, Hackl C, Kennel R, Rodriguez J (2019) Sensorless predictive speed control
of permanent-magnet synchronous generators in wind turbine applications. In PCIM Europe,
International Exhibition and Conference for Power Electronics, Intelligent Motion, Renewable
Energy and Energy Management (1–8) VDE
8. Ouafia F, Ahmed A (2020) A direct power control of the PWM rectifier for SEIG feeding
resistive load in wind energy systems. In 5th International Conference on Renewable Energies
for Developing Countries, pp 1–6. IEEE
9. Mohan H, Pathak MK, Dwivedi SK (2019) Direct power control of induction motor drives.
In IEEE 13th International Conference on Compatibility, Power Electronics and Power
Engineering pp 1–5
10. Boudries ZOUBIR, Aberbour A, Idjdarene KASSA (2016) Study on sliding mode virtual
flux-oriented control for three-phase PWM rectifiers. Rev Roum Sci Techn–Électrotechn
Énerg 61(2):153–158
11. Xiong L, Li P, Wang J (2020) High-order sliding mode control of DFIG under unbalanced grid
voltage conditions. Int J Electr Power Energy Syst 117:105608
Improved DTC of the PID Controller
by Using Genetic Algorithm of a Doubly
Fed Induction Motor
Said Mahfoud, Aziz Derouich, Najib El Ouanjli, Mohammed Taoussi,
and Mohammed El Mahfoud
Abstract This paper presents a new strategy for the Direct Torque Control (DTC) of
the Doubly Fed Induction Motor (DFIM) powered by two voltage inverters. The speed
regulation of this control is done by a PID (proportional integral derivative) controller
optimized by a Genetic Algorithm (GA). The classic DTC has several advan-
tages such as (dynamic, robustness, ease of implementation, high performance…).
However, it has ripples of the electromagnetic torque and variable switching frequen-
cies that generate vibrations of the machine causing rapid aging. To overcome the
disadvantages of classic DTC control, the new GA-DTC control strategy based on the
PID controller optimized by the genetic algorithm, have the tendency to operate the
system throughout its operating range. The entire system is implemented and vali-
dated on the Matlab/Simulink environment to improve machine performance and
behavior.
1 Introduction
During the mid-80s, the control technique of the induction motor, known as Direct
Torque Control or DTC, has appeared to compete with traditional techniques. This
strategy has been introduced by TAKAHASHI [1] and DEPENBROCK [2]. Its prin-
ciple for a DFIM is based on a direct identification of the control pulses applied to
the voltage inverters switches, in order to preserve the stator and rotor flux and elec-
tromagnetic torque within two predefined hysteresis bands. Such an application of
this technique makes it possible to ensure decoupling between the flux and the torque
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1687
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_153
1688 S. Mahfoud et al.
control without using the Pulse Width Modulation (PWM) or coordinate transforma-
tion. The important features of DTC are its fast dynamic torque response, robustness,
ease of implementation, and high performance [3, 4]. However, this strategy has two
important disadvantages: First of all, the frequency of switching is highly variable.
Second of all, the torque and flux ripples remains poorly handled in the entire oper-
ating range. It is worthwhile to note that the torque ripples are making additional
noises and vibrations. Therefore, they are causing fatigue in the machine shaft. In
order to reduce the impact of these phenomena on machine life, it is estimated that
intelligent techniques can provide improvements.
Currently, many researchers have proposed solutions to improve the performance
of classic DTC control, which are based on artificial intelligence. These approaches
are very robust to measure noise and parametric variations.
In [5–7], the authors have proposed many techniques to enhance the efficiency
of direct torque control using intelligent techniques. They are called Direct Torque
Fuzzy Control (DTFC), Direct Torque Neural Control (DTNC), and Direct Torque
Neural-Fuzzy Control (DTNFC). They use fuzzy logic controllers or neural networks
or an Adaptive Neuro-Fuzzy Inference System (ANFIS) combining artificial neural
networks and fuzzy logic, which replace truth tables and hysteresis comparators, to
generate the voltage vector in order to direct flux and torque to their references.
In [9] the authors have proposed a new DTC strategy of the DFIM, The PI speed
controller is optimized by using the genetic algorithm of DTC control at the stator, in
order to minimize torque ripples. But this method does not allow us to take advantage
of the double speed band of DFIM [10].
In order to solve this problem, this work is articulated on the application of the
DTC control on both sides of the DFIM and the optimization of PID controller
parameters based on genetic algorithm to resolve the precedent problems. In this
article, the contributions are focused on the following objectives:
• Minimization of the flux and torque ripples
• Reduction of response time and speed rejection time
• Minimization of the speed and torque overshoot
This article is organized according to the following form: Sect. 2 Presents the
mathematical model of the DFIM, Sect. 3 Describe the operation principal and math-
ematical model of the DTC, Sect. 4 is focused to study the genetic algorithm, Sect. 5
is articulated for interpreted the results of simulation and finally the conclusion of
control results and suggestion of future works.
The mathematical model of doubly fed induction motor appropriate for studying
the behavior of direct torque control is the model of two-phase represented by the
coordinates (α, β), it is represented by the following equations [11]:
Improved DTC of the PID Controller by Using Genetic Algorithm ... 1689
• Electrical equations:
⎧
⎪
⎪
dψsα
⎪ vsα = Rs .i sα +
⎪
⎪
⎪ dt
⎪
⎪
⎪
⎪ dψ sβ
⎨ vsβ = Rs .i sβ +
dt (1)
⎪
⎪ dψr α
⎪ vr α = Rr .ir α +
⎪ + ωm .ψrβ
⎪
⎪ dt
⎪
⎪
⎪
⎪ dψrβ
⎩ vrβ = Rr .irβ + − ωm .ψr α
dt
• Magnetic equations:
⎧
⎪
⎪ ψ = L s i sα + L m .ir α
⎨ sα
ψsβ = L s i sβ + L m .irβ
(2)
⎪
⎪ ψ = L r ir α + L m .i sα
⎩ rα
ψrβ = L r irβ + L m .i sβ
• Mechanical equations:
⎧
⎨ Tem = p.(ψsα i sβ − ψsβ i sα )
(3)
⎩ J. d + f. = Tem − Tr
dt
The DTC is less sensible in front of parametric variations of the motor and it allows
us to obtain precise and fast torque response. The objective of this strategy is to
control directly the machine torque and flux. In this context, a hysteresis comparator
is used to compare the estimated value and the reference one, and then the state of
the inverters is controlled by a predetermined selection table. Figure 1 illustrates the
structure of the DTC control strategy applied to the doubly fed induction motor.
In the fixed (α, β) plane associated with the stator and rotor. The rotor and stator flux
are estimated according to the equations below [11].
t
ψ s (t) = 0 (V s + Rs .I s ).dt
t (4)
ψ r (t) = 0 (V r + Rr .I r ).dt
1690 S. Mahfoud et al.
Fig. 1 Structure of the direct torque control strategy optimized by the Genetic Algorithm
The flux are defined by their modules and their positions, which are given by the
following relationship:
ψ̂ = ψ̂α2 +ψ̂β2
ψ̂
(5)
θ = ar ctg( ψ̂β )
α
By using the flux in the reference (α, β), the electromagnetic torque can be
calculated using a mathematical equation.
Fig. 3 Hysteresis comparators (a) of the three-levels torque (b) of the two-levels flux
Two hysteresis comparators with two-levels, which are easy to use. Their roles are to
maintain the end of each vector of the flux in a circular crown as shown in Fig. 2, and a
three-levels hysteresis comparator Fig. (3, a), allows controlling the electromagnetic
torque developed by the motor in both directions of rotation, they are generating
either a positive torque or a negative one.
Depending on the sector and the evolution of the torque and flux, you should select
the voltage vector to be applied in order to respect the references of the torque and
flux. The truth table for selecting the appropriate vector is shown in Table 1.
1692 S. Mahfoud et al.
Genetic algorithm is a research strategy to provide the best solution for non-linear
systems. Classic DTC controller and machine are affected by variable parameters,
which makes the system non-linear, for this situation, a genetic algorithm generates
the optimal values for KP , KI and KD of the PID controller at each occurrence of a
change, in the system in order to master the situation. Figure 4 illustrates the reduced
structure of the system.
The genetic algorithm belongs to the category of evolutionary algorithm, which
uses techniques based on evolutionary biology, such as selection, crossover, and
mutation [12], The sequences of operations involved in GA are described in Fig. 5,
which presents a flowchart that respects the evolutionary rules of a genetic algorithm.
The implementation of the GA-DTC control and the DFIM is realized in the
Matlab/Simulink platform. The configuration of the global system is done according
to the parameters Tables 5, 6 and 7. To find out the behavior of the system, the latter
is subjected to variable references of speed and torque to put it in situations similar
to reality. The results of the simulation for both controls (GA-DTC and classic DTC)
were tested using a 1.5 kW machine.
Improved DTC of the PID Controller by Using Genetic Algorithm ... 1693
Figures (6, a), (6, b), (6, c) and (6, d), respectively show the variation of the speed,
and the variation of the electromagnetic torque and the stator and rotor flux, which
are obtained without load as a condition of starting and at nominal load from 0.5 s
as normal operating condition.
Figures (6, a), and (6, b) illustrate, the speed and torque characteristics of a doubly
fed induction motor controlled by the GA-DTC, which quickly reach their references
without overshoot compared with a classic DTC during no-load starting and during
nominal load starting.
For the speed characteristic the response time is minimized by a rate greater than
37% (48.5 ms for the classic DTC and 30.4 ms for the proposed GA-DTC), and the
overshoot is reduced by approximately 100% (9 rad/s for classic DTC and 0 rad/s for
proposed GA-DTC control), and the rejection time is reduced by 16.73% (23.3 rad/s
for the classic DTC and 19.4 rad/s for GA-DTC control).
For the characteristic of the electromagnetic torque, the ripples are reduced by
more than 22% (4,264 Nm for classic DTC and 3,297 Nm for the proposed GA-DTC).
1694 S. Mahfoud et al.
For the stator and rotor flux characteristics, the ripples are reduced to 36.05% and
23.74% respectively (0.1079 Wb and 0.069 Wb for the classic DTC and 0.0514 Wb
and 0.0392 Wb for the proposed GA-DTC).
Improved DTC of the PID Controller by Using Genetic Algorithm ... 1695
The results presented in Table 2 showed that the optimization of the DTC
control PID controller by the genetic algorithm improved the classic DTC control
performances (dynamics, robustness…).
6 Conclusion
In this article, the optimization of the PID controller of the classic DTC control is
achieved by a genetic algorithm, which has been applied to a doubly fed induction
motor powered through two inverters. The modeling and details of the system are
introduced and implemented in the Matlab/Simulink platform. The results of the
simulation improve the performance of the doubly fed induction motor, such as
minimizing the response time of 37.32% and the rejection time is improved by
16.73%, and the torque ripples which are minimized by up to 22.68% . The proposed
control has increased the performance of the classic DTC control at the transient and
dynamic regime. In order to make a progress of the research the next work will be
on the implementation of the proposed control on the dSPACE DS1104 board to test
the validity of this technique.
Appendix
Table 3 Nomenclature
Parameters Description
Vsα , Vsβ , Vrα and Vrβ Stator and rotor voltages in (α, β) plan
Udcs and Udcr Stator and rotor directs voltages
Isα , Isβ , Irα and Irβ Stator and rotor currents in (α, β) plan
sα , sβ , rα and rβ Stator and rotor flux in (α, β) plan
Rs , Rr Stator and rotor resistors
Ls , Lr Stator and rotor inductors
Lm Mutual Inductance
P Number of pairs of poles
ωr Rotor angular speed
ωs Stator angular speed
Rotation speed
Tem Electromagnetic torque
Tr Resistant torque
f Viscous friction coefficient
Table 4 Abbreviations
Abbreviation Wording
DFIM Doubly Fed Induction Motor
DTC Direct Torque Control
GA Genetic Algorithm
GA-DTC Genetic Algorithm-Direct Torque Control
PID Proportional Integrator Derivator
DTFC Direct Torque Fuzzy Control
DTNC Direct Torque Neural Control
DTNFC Direct Torque Neural-Fuzzy Control
ANFIS Adaptive Neuro-Fuzzy Inference System
Table 5 (continued)
Symbols Values (unit)
Rr 1.68
Ls 0.295 H
Lr 0.104 H
Lm 0.165 H
f 0.0027 kg.m2 /s
J 0.001 kg.m2
References
10. Zemmit A, Messalti S, Harrag A (2017) A new improved DTC of doubly fed induction machine
using GA-based PI controller. Ain Shams Eng J 9:1877–1885
11. El Ouanjli N, Derouich A, El Ghzizal A et al (2019) Modern improvement techniques of direct
torque control for induction motor drives - a review. Prot Control Mod Power Syst 4:11. https://
doi.org/10.1186/s41601-019-0125-5
12. Goldberg DE (1989) Sizing populations for serial and parallel genetic algorithms. In:
Proceeding of the third international conference on genetic algorithms, CL, USA
State Feedback Control of DC-DC
Converter Using LQR Integral
Controller and Kalman Filter Observer
Djamel Taibi , Toufik Amieur , Mohcene Bechouat , Sami Kahla ,
and Moussa Sedraoui
Abstract In this paper, the linear state feedback control using LQR controller for
a DC/DC converter in the case of negative voltages topology is presented in order
to achieve a particular desired behavior. To guarantee a zero steady-state error, we
introduce an integral action, which will work out this problem by assuring that
the steady-state error will end up to zero. For filtering and state estimation with
a low cost and less complexity a state observer is obtained based a Kalman Filter
observer. Detailed simulation study is presented to demonstrating the robustness and
effectiveness of the proposed control scheme.
1 Introduction
Power supply technology enables technologies that allow us to operate circuits and
electronic systems. All digital and analog circuits require a power source to operate.
Many of these circuits require multiple DC supply voltages. DC power supplies
D. Taibi (B)
Département de Génie Électrique, Université de Kasdi Merbah, Ouargla, Algeria
T. Amieur
Département de Génie Électrique, Université de Larbi Tebessi, Tebessa, Algeria
e-mail: [email protected]
T. Amieur · M. Bechouat · M. Sedraoui
Laboratoires des Télécommunications LT, Université 8 mai 1945, Guelma, Algeria
e-mail: [email protected]
M. Bechouat
Faculté des Sciences et Technologie, Université de Ghardaïa, Noumirat BP 455,
Route Ouargla Ghardaia, 47000 Ghardaia, Algeria
S. Kahla
Centre de Recherche en Technologies Industrielles CRTI, 16014 Cheraga, Alger, Algeria
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1699
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_154
1700 D. Taibi et al.
By applying Kirchhoff’s laws in the two previous circuits, we get the following
dynamics:
In the switching ON, the following dynamic is obtained [12]:
⎧ di
⎪
⎨L =E
dt (1)
⎩ C dv = − v
⎪
dt R
In the switching OFF, the following dynamic is obtained [12]:
⎧ di
⎪
⎨L =v
dt (2)
⎩ C dv = −i − v
⎪
dt R
From the two Eqs. (1) and (2), we can obtain a single unified model, which is
u ∈ {0, 1}.The dynamics of the converter are therefore as follows [13]:
L di
dt
= (1 − u)v + u E
(3)
C dv
dt
= −(1 − u)i − Rv
1702 D. Taibi et al.
3 LQR Control
The Eq. (4), describe the state space model of a system: [14]
ẋ(t) = Ax(t) + Bu(t)
(4)
y(t) = C x(t) + Du(t)
To define the optimal control inputs and optimize the state variables, the following
cost function J must be minimized [15]:
∞
J= x T Qx + u T Ru dt (6)
0
Q and R two positive definite matrix, Q and R are chosen, the LQR control
problem reduces to finding K that minimizes (6).
Solution P of Riccati equation [15]:
P A + A P + Q − P B R −1 B P = 0 (7)
The gain matrix of the optimal control vector can be calculated using the following
equation [15]:
K = R −1 B P (8)
The schematic diagram LQR with integral is shown in Fig. 4 [10, 16].
The LQR with integral action require the following extended matrix configuration:
A 0 B
 = , B̂ = , Ĉ = 0 0 , K̂ = k −ki
−C 0 0
State Feedback Control of DC-DC Converter ... 1703
With
k: State-feedback vector gain.
ki : Integral action.
The cost function is known by the following relationship [15]:
∞ T
J= e (t)Qe(t) + u T (t)Ru(t) dt
0
 T P + P  − P B̂ R −1 B̂ T P + Q = 0 (11)
K̂ = R −1 B̂ T P (12)
4 Kalmen Filter
We consider that only the output voltage is sensed. Then, to estimate the inductor
current and reduce the noisy environment effects, a Kalman filter is introduced [10]:
ˆ = A x̂(t) + Bu(t) + ω(t)
ẋ(t)
(13)
ŷ(t) = C x̂(t) + ν(t)
1704 D. Taibi et al.
where ω(t) is a process noises, and ν(t) is the voltage sensor noise.
The state equations of the Kalman filter can be carried out as follows [17]:
d ẋˆ = A(t)ẋ(t)dt
ˆ ˆ
+ B(t)u(t)dt + K K (t)(y(t)dt − C(t)ẋ(t)dt) (15)
d
PK (t) = A(t)PK (t) + PK (t)A T (t) − PK (t)C(t)T R −1
K (t)C(t)PK (t) + Q K (t)
dt
(17)
x̂(t0 ) = x0 , P(t0 ) = P0
5 Simulation Results
Open-Loop Open-Loop
Closed- Closed-
In this case, the simulation of the buck boost converter is carried out for two cases
for the control behavior of the converter in the Open and Closed-Loop control using
LQR. The SimPowerSystems model is shown in Fig. 5.
According to results of the simulation, the waveform of output voltage and the
inductance current have the same waveform characteristic for the open and closed
loop control (Fig. 6).
The value of the output voltage in the closed-loop is approximately −40 V and
about −37 V in the open-loop. Then the average value of the inductance current is
about 6 A and 5.6 A in closed and open-loop respectively. A comparison between
the simulation results shows that the control with closed-loop using LQR has better
dynamics response compared with open-loop as well as in the maximum peak
overshoot/ undershoots, settling time.
1706 D. Taibi et al.
Fig. 7 SimPowerSystems model of buck-boost converter with LQR and Kalman filter observer
Estimated
Measured
Estimated
Measured
The estimated and measured output voltage and inductance current at an output
voltage reference of −40 V are shown in Fig. 8 and Fig. 9. In both static and dynamic
cases, it can be seen that the signals estimated by the KF observer track the real signals
very well.
In many applications, the measurement noise caused by the hardware or the environ-
ment has a significant impact on the system. The estimation accuracy of KF is tested
in this paper under noisy output voltage measurement. Figure 10 shows the noise
injected at the output voltage in the range of [0–2 A]. The noise is zero mean, white
Gaussian. The aim of the voltage injection is to observe the low pass filter characteris-
tics of KF. Figure 11 shows the output voltage, the inductor current shown in Fig. 12.
The accuracy of the state estimate can be increased by increasing the covariance of
Noisey
Estimated
Measured
Estimated
Measured
the measurement noise under noisy conditions, which makes the system model will
have more importance.
6 Conclusion
This paper has demonstrated the process of applying the state feedback control of a
buck-boost converter using LQR with integral action controller. The clear advantages
of state feedback, that has a positive effect on response settling time, reducing the
undesirable peak overshoots and serve having a less oscillated performance, referring
that this approach doesn’t provide a zero static error, this latter has been solved by
adding an integral action to state feedback control. The simulation results found that
adding a Kalman filter reduces the cost and impact of the sensor on LQR tracking
performance.
State Feedback Control of DC-DC Converter ... 1709
References
1 Introduction
In the world as it is today and without electricity, daily life would be difficult to
imagine. It is, therefore, necessary to produce electric energy efficiently and contin-
uously. Following the terrible damage caused by the production of electricity on
the environment employing fossils combustible such as coal, oil, and natural gas.
Renewable energy-based on inexhaustible energy sources remains the best solution
to replace the existing energy sources [1]. Recently, wind energy, which is consid-
ered clean and sustainable energy, has known a great development and exploitation,
thanks to the strong growth in the research domain. The wind energy system has
several configurations. Figure 1 represents the studied system configuration, which
is widely used and dedicated to wind energy applications. The wind turbine shaft is
connected to the doubly fed-induction generator through the gearbox to conform the
slow-speed to the high one. DFIG stator is directly coupled to the electrical network.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1711
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_155
1712 E. Chetouani et al.
However, the rotor is branched to the power grid via bidirectional converters. Using
the generator integrated with a wind energy system has many advantages such as the
ability to use a partial sized converter in the rotor to control the power, minimizing
power losses and cost, making fewer efforts on mechanical parts, noise reduction, the
control of active and reactive power, and a controllable power factor [2]. The gener-
ator is a non-linear system and has complex and coupling equations. The vector
control based on the stator or rotor field orientation combined with the Park trans-
formation makes the control of the generator a simple and easier task. Besides, this
vector control permits the command of the DFIG as a separately excited DC machine,
which makes it very popular in the industry. The direct and indirect methods are two
variants of the vector control. These types have been compared by Bouderbala et al.
[3]. The authors have shown that indirect control presents a satisfactory perfor-
mance than direct vector controls. The vector command based on the classical PI
controller has demonstrated performance limitations, especially when the parame-
ters of the generator change. Besides, the non-linear Backstepping approach, based
on the stability of the Lyapunov function, has been proposed by several researchers
to improve the performance of the system. S. Mensou et al. [4] have established
a Backstepping strategy to control the whole wind energy conversion system. The
produced mechanical power from wind energy depends on the characteristics of each
turbine, and eventually the wind variable speed. Consequently, tracking the maximal
power generated is required when the profile airspeed changes. To perform the MPT,
several techniques have been developed, and they can be divided into two categories
[5, 6]. The first one has required the characteristic aerodynamic curve of wind speed
information, and the second one does not need any information about wind speed
to generate excellent speed reference. The objective of this paper is to establish
a comparative study between the Backstepping and the Open Loop Indirect Field
Backstepping and Indirect Vector Control for Rotor Side Converter ... 1713
Oriented strategy for piloting the RSC converter of the induction generator. Also, the
maximum power following with speed control is established to generate the power
reference and the optimal rotation speed driving the generator. The article is organized
in this way: Sect. 2 introduces the modeling of the WECS. Section 3 discusses the
MPT with a speed control strategy computed by the classical PI controller. Section 4
proposes the control of the stator powers (Active and Reactive) by using Backstep-
ping and indirect vector control strategies. Section 5 examines and compares the
simulation results. Finally, Sect. 6 highlights the conclusion.
1
PTu = · Cp (λ, β) · ρ.π · R2 · V3W (1)
2
The mechanical torque TTu is written as below:
PTu
TTu = (2)
Tu
Where, ρ is the air density (kg/m3 ), R is the blade ray (m), Tu is the angular
velocity of the turbine, and CP represents the performance factor of the turbine. CP
can be formulated as [8]:
π(λ + 0.1)
CP (β, λ) =[0.5 − 0.0167.(β − 2)].sin
18.5 − 0.3.(β − 2)
− 0.00184.(λ − 3).(β − 2) (3)
This coefficient is calculated by the tip speed λ and the angle β of the blade pitch.
The latter is fixed to β = 2◦ for having CPmax . Equation (4) defines the expression of
λ [7]:
R.Tu
λ= (4)
V
1714 E. Chetouani et al.
Equation (5) presents the mechanical equation of the system, taking into consider-
ation that the overall mechanical dynamics are brought back to the turbine shaft,
[5–7]:
dmec
Jtot . + f.mec = Tg − Tem (5)
dt
Where
TTu mec
Tg = and GB = (6)
GB Tu
Where Jtot is the overall inertia of wind energy system, TTu is the turbine torque,
TTem is the electro-magnetic torque of the generator, f is the overall viscous coefficient
of friction, mec is the rotational speed at the rotor shaft of the gearbox (rad/s) and
GB is the gearbox multiplier.
The Park transformation allows simplifying the general electrical model, which is
determined by the following equations. Stator and rotor voltages equations are [2,
8]:
⎧
⎪
⎪ V = Rs isd + dϕdtsd − ωs ϕsq
⎪ sd
⎨ dϕ
Vsq = Rs isq + dtsq + ωs ϕsd
(7)
⎪ dϕrd
⎪ Vrd = Rr ird + dt − ωr ϕrq
⎪
⎩ V = R i + dϕrq + ω ϕ
rq r rq dt r rd
Where, Vr and Vs are the voltages; is and ir are the currents; ϕs and ϕr are the
flux; Rs and Rr are the resistances; ωs and ωr are the angular frequencies; Ls and Lr
are the inductances; M is the mutual inductance. The r and s denote the rotor and
stator, respectively. The electromagnetic torque can be expressed as follows [8]:
M
Tem = −p. (irq ϕsd − ird ϕsq (8)
Ls
During the normal functioning of the wind turbine, the maximum power control
method is developed to exploit the energy available in the wind as much as possible.
The MPT method with mechanical speed control is established. This technique
consists of maintaining the generator speed at its reference, which is maximized
when the Cp is optimal. The electro-magnetic torque developed by DFIG is equal to
its reference value imposed by the control defined as:
The optimal electro-magnetic torque Tem-opt for obtaining a rotation speed equal
to the optimal speed is given as follows:
1
Tem−opt = Kpmppt + Kimppt . .[mec−opt − mec ] (12)
S
Where, KPmppt and Kimppt are the PI controller gains. The optimal speed (mec-opt )
is:
Vw .λopt
mec−opt = GB .Tu−opt ; With Tu−opt = (13)
R
Determination of the PI Gains for MPPT. The PI controller parameters are determined
by the pole compensation method. The time constant of the system (Tsys ) is:
Jtot
Tsys = (14)
f
The gains of the controller are expressed as:
To command independently the powers of generator and to deal with the coupling
problem of the system, the indirect vector control is applied. In this paper, the powers
are controlled in the open loop and the rotor currents are controller in the closed-
loop. Stator flux is considered constant and is oriented according to d-axis. The stator
resistance is neglected and the stator voltage equation can be simplified as [2, 8]:
From Eqs. (17)–(18), the rotor currents expressions are deduced as follows:
M2 M2
ird = Vrd + gωs Lr − irq / Rr + Lr − .s (19)
Ls Ls
M2 Vs M M2
irq = [V rq − gωs Lr − ird − g ]/ Rr + Lr − .s (20)
Ls Ls Ls
∗ Ls
irq =− .P∗ (21)
M.Vs s
∗ Ls Vs2
ird =− .(Qs∗ − ) (22)
M.Vs Ws Ls
The voltage references are expressed easily from Eqs. (17)–(18) as follows:
1 Vs M
∗ ∗
Vrq = irq − irq .[Kp−rsc1 + Ki−rsc1 . ] + erd + g (23)
S Ls
∗
∗
1
Vrd = ird − ird .[Kp−rsc2 + Ki−rsc2 . ] + erq (24)
S
Backstepping and Indirect Vector Control for Rotor Side Converter ... 1717
Where:
M2 M2
erd = gωs Lr − .ird And erq = gωs Lr − .irq (25)
Ls Ls
M2
Ts = (Lr − )/Rr (26)
Ls
1 M2 Kprsc .Rr Ts
Kprsc = .(Lr − ) And Kirsc = 2 With Trsc = (27)
Trsc Ls (Lr − Ls )
M 100
The RSC is designed by Backstepping procedure to pilot the powers. For this task,
virtual control should be designed based on the rotor currents. The system stability
and performance are attained by utilizing a Lyapunov theorem, which is exploited to
establish the control of the system. Considering, the Eqs. (19)–(20), the derivation
of the rotor currents equations can be deduced as follows:
dird 1 M2
= (Vrd − Rr ird − g.ωs Lr 1 − .irq ) (28)
dt Lr . 1 − M2 Lr .Ls
Lr .Ls
dirq 1 M2 M.Vs .g
= (Vrq − Rr irq − g.ωs Lr 1 − .irq − ) (29)
dt Lr . 1 − M2 Lr .Ls Ls
Lr .Ls
Calculating the errors between desired rotor currents and the actual one to stabilize
it by Backstepping in the first step as:
∗ ∗
εrsc1 = irq − irq And εrsc2 = ird − ird (30)
The Lyapunov candidate function is defined, in a second step, as follows [4, 9]:
ε2rsc1 ε2
Lf−rsc = + rsc2 (34)
2 2
The derivation of Lyapunov function is expressed as:
So, replacing the errors derivation in Eq. (35), the Lyapunov derivative rewrites
as:
⎡ ⎤
∗
⎣
dirq 1 M2 M.Vs .g ⎦
L̇f−rsc =εrsc1 . − (Vrq − Rr irq − g.ωs Lr 1 − .irq − )
dt 2
Lr . 1 − LM.L Lr .Ls Ls
r s
⎡ ⎤
∗ 2
⎢ di 1 M ⎥
+ εrsc2 .⎣ rd − (Vrd − R ird + g.ωs Lr 1 − .irq )⎦ (36)
dt L . 1− M
2 Lr .Ls
r Lr .Ls r
Where, K1 and K2 are positive constants. By equalizing the Eq. (36) and the
Eq. (37), the virtual commands Vrd and Vrq are deduced directly as follows:
⎡ 2
⎤
∗
dirq Rr g.ωs Lr 1 − LM.L
⎢ + irq + .ird ⎥
r s
⎢ 2 2 ⎥
⎢ M2
dt Lr . 1 − LM.L Lr . 1 − LM.L ⎥
∗ =L . 1− ⎢ r s r s ⎥
Vrq r .⎢ ⎥ (38)
Lr .Ls ⎢ M.Vs .g ⎥
⎢ Ls ⎥
⎣ + + K .ε
1 rsc1 ⎦
2
Lr . 1 − LM.L
r s
Backstepping and Indirect Vector Control for Rotor Side Converter ... 1719
⎡ 2
⎤
M 2 di ∗
R g.ωs Lr 1 − LM.L
∗ =L . 1−
Vrd .⎣ rd +
r
ird −
r
.irq + K2 .εrsc2 ⎦
s
r
Lr .Ls dt 2
Lr . 1 − LM.L
2
Lr . 1 − LM.L
r s r s
(39)
With help of the before-mentioned mathematical models, the wind chain is modeled
and simulated employing the parameters presented in Tables 1 and 2 under
Matlab/Simulink software. The wind profile used in this work is presented in Fig. 2.
However, Fig. 3 illustrates the produced mechanical power, which is pursuing the
evolution of the wind speed. From 0.5 (s) to 0.8 (s) the power is equal to 5 MW, which
corresponds to the rated value of the wind velocity that is 12.5 (m/s). However, Fig. 4
shows the mechanical speed computed by the MPT strategy based on PI controller.
This power will be utilized as a reference of power for driving the system. Figures 5
and 6 show the efficiency of the MPT strategy for making constant the tip speed ratio
at its best value, which is λopt = 9.19 and the coefficient of power is preserved at
its maximum (Cpmax = 0.5). So, the maximal power is achieved. As you can see in
Fig. 7, the stator active power is computed by the Backstepping (BS), in which the
positive constants are chosen by trial and error method as follows: K1 = K2 = 9000,
to satisfy convergence condition, and the Indirect Field-Oriented control based on the
6 Conclusion
In this paper, the wind energy conversion system is modeled and simulated based
on variable wind speed. The Maximum Power Tracking with speed regulation is
established. Then, the control of the rotor side converter is designed by employing
the IFOC based on the classical PI controller, in which the gains are calculated
by the pole compensation method, and the Backstepping based on the Lyapunov
stability which the positive constants are chosen. The results of the two strategies
are compared and analyzed. It is observed that the Backstepping brings satisfactory
and attractive results than the indirect vector control under wind speed variation. The
robustness test shows that the proposed method ensures good stability of the system
when the inductance of the DFIG changes. The Backstepping technique improves
the stability, the time response, and ameliorates the static error than indirect vector
control based on the classical PI controller.
Appendix
References
1. Noussi K, Abouloifa A, Katir H, Lachkar I (2019) Modeling and control of a wind turbine
based on a doubly fed induction generator. In: 2019 4th world conference on complex systems
(WCCS), Ouarzazate, Morocco, pp 1–5. https://doi.org/10.1109/ICoCS.2019.8930738
2. Lamnadi M, Trihi M, Bossoufi B, Boulezhar A (2016) Modeling and control of a doubly-
fed induction generator for wind turbine- generator systems. Int J Power Electron Drive Syst
(IJPEDS). 7(3):973–985. https://doi.org/10.11591/ijpeds.v7.i3.pp982-994
3. Bouderbala M, Bossoufi B, Lagrioui A, Taoussi M, Aroussi H, Ihedrane Y (2018) Direct and
indirect vector control of a doubly fed induction generator based in a wind energy conversion
system. Int J Electr Comput Eng (IJECE). 9(3):1531–1540. https://doi.org/10.11591/ijece.v9i3.
pp1531-1540
4. Mensou S, Essadki A, Minka I, Nasser T, Bououlid Idrissi B (2018) Backstepping controller
for a variable wind speed energy conversion system based on a DFIG. Int J Electr Comput Eng
12(9)
5. Omar Baba A, Liu G, Chen X (2020) Classification and evaluation review of maximum
power point tracking methods. Sustainable Futures 2:100020. https://doi.org/10.1016/j.sftr.2020.
100020
6. Errami Y, Ouassaid M, Cherkaoui M, Maaroufi M (2015) Maximum power point tracking control
based on a nonlinear backstepping approach for a permanent magnet synchronous generator wind
energy conversion system connected to a utility grid. Energy Technol 3:743–757. https://doi.
org/10.1002/ente.201500026
7. Errami Y, Obbadi A, Sahnoun S, Benhmida M, Ouassaid M, Maaroufi M (2016) Design of a
nonlinear backstepping control strategy of grid interconnected wind power system based PMSG.
In: AIP conference proceedings, vol 1758, p 030053. https://doi.org/10.1063/1.4959449
8. Bouderbala M, Bossoufi B, Lagrioui A, Taoussi M, Alami Aroussi H, Ihedrane Y (2018) Direct
and indirect vector control of a doubly-fed induction generator based in a wind energy conversion
system. Int J Electr Comput Eng (IJECE) 9(3):1531–1540. https://doi.org/10.11591/ijece.v9i3.
pp1531-1540
9. Nadour M, Essadki A, Nasser T (2017) Comparative analysis between PI & backstepping control
strategies of DFIG driven by wind turbine. Int J Renew Energy Res 7(3):1307–1316
Integral Backstepping Power Control
of DFIG Based Nonlinear Modeling
Using Voltage Oriented Control
Abstract In this paper, the control of the wind system, the mechanical part consisting
of the wind turbine with maximum power point tracking control, and the electrical
part using a double fed induction generator is studied, taking into consideration the
full nonlinear model of the DFIG with voltage oriented control. The control strategy
is applied to the rotor side converter with an integral backstepping controller of the
active and reactive stator powers operating at sub-synchronous speed. The simulation
results show a good capability to monitor the approach proposed at different variable
wind speeds.
1 Introduction
The wind has many advantages and has been used by man for many thousands of
years for both sailing and milling. Today, it is one of the main uses of wind energy,
because of its free availability, inexhaustible and its clean energy [1, 2].
DFIG is broadly applied in the Wind Energy Conversion System (WECS) due to
its mature technology and high power production efficiency. With the advantages of
long distance and large transmission capacity, series compensated transmission line
is often used to transfer large capacity wind energy to remote load center.
Backstepping control uses the Lyapunov function to ensure convergence of the
error of position tracking from all available initial conditions [3, 4]. An integral added
is integrated in the reverse design. The error integral is added to the first function
of stabilization in the backstepping approach design to remove the steady state error
and increase the rejection of disturbances [5].
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1725
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_156
1726 M. Loucif et al.
The structure of the paper is as follows. Section 1 discusses the modeling and
control of the wind turbine to get the maximum power point tracking with linear
controller PI. Section 2 describes a complete model of the double fed induction
generator within the framework of synchronous rotation dq and a new modeling
approach that uses active and reactive power as states. We adopt the voltage oriented
approach for DFIG to propose a simplified active and reactive stator power control.
The integral backstepping design is discussed in Sect. 3. Section 4 and 5 provide the
numerical results and concluding remarks, respectively.
The power and the torque delivered by the wind turbine are defined by [6, 7]:
⎧
⎪
⎪ 1
⎨ Paer = C p (λ, β).ρ.π.R .V
2 3
2
(1)
⎪
⎪ ρ 1
⎩ Taer = C p (λ, β). .π.R 2 .V 3
2 tur b
With ρ is the air density, V is the wind speed; R is the rotor radius.
The expression of power coefficient Cp(λ, β) who is a function of pitch angle β
and tip speed ratio λ, is written as shown below [8]:
π.(λ + 0.1)
C p = 0, 5176 − 116.(β − 0, 4). sin − 5.(λ − 21).(β − 0, 4)
14.8 − 0.3.(β − 2)
(2)
tur b .R
λ= (3)
V
The expression of the gearbox G is given by:
Ttur b
G= = (4)
Tg tur b
d
J. = Tg − Tem − B. (5)
dt
The power coefficient expression (2) has been shown to have a unique maximum,
corresponding to the optimal choice of the tip speed ratio λopt = 9,15 and the power
coefficient take a maximum value of C pmax = 0,5 for a pitch angle is fixed β = 2°
[9].
Integral Backstepping Power Control of DFIG Based Nonlinear Modeling ... 1727
To get the maximum power, we use the MPPT control with wind speed measure-
ment developed in paper [6]. For the optimal tip speed ratio λopt , (3) and (4) provide
the optimal rotational speed generator as
λopt .V
∗ = (6)
R
The mathematical model of DFIG has been obtained after the according to the
voltage equations and equations of flux, expressed in a dq reference frame rotating
at synchronous speed, the differential equations of the stator and rotor circuits of the
generator with stator and rotor current as state variables can be obtained by [10, 13].
⎧
⎪
⎪
di sd
= −a1 .i sd + (a.ω + ωs ).i sq + a3 .ir d + a5 .ω.irq + b1 .Vsd − b3 .Vr d
⎪
⎪
⎪
⎪ dt
⎪
⎪
⎪
⎪ di sq
⎨ = −(a.ω + ωs ).i sd − a1 .i sq − a5 .ω.ir d + a3 .irq + b1 .Vsq − b3 .Vrq
dt
⎪ ω
(7)
⎪ dir d
⎪ = a4 .i sd − a5 .ω.i sq − a2 .ir d + ωs − .irq − b3 .Vsd + b2 .Vr d
⎪
⎪ σ
⎪
⎪ dt
⎪
⎪ ω
⎪ di
⎩ rq = a6 .ω.i sd + a4 .i sq − ωs − .ir d − a2 .irq − b3 .Vsq + b2 .Vrq
dt σ
M2 Rr .Msr Rs .Msr
Where, σ = 1 − L s .Lsr r , a = 1−σ σ
, a1 = σ.LRs
s
, a2 = σ.L Rr
r
, a3 = σ.L s .L r
, a4 = σ.L s .L r
,
a5 = σ.L s , a6 = σ.L r , b1 = σ.L s , b2 = σ.L r , b3 = σ.L s .L r .
Msr Msr 1 1 Msr
Stator and rotor resistances (Rs and Rr ). ω = p. electrical speed and p is
the number of poles pair. Stator and rotor inductances (L s L r ) and the magnetizing
inductance (M sr ). The direct and quadrature stator and rotor currents and voltage are
given as follow (isd , isq , ird , irq and V sd , V sq , V rd , V rq ).
The electromagnetic torque can be described by fundamental torque Eq. (8):
⎧
⎪
⎪ Msr
⎨ Te = p ϕsq .ir d − ϕsd .irq
Ls
(8)
⎪
⎪ d 1
⎩ = (Te − Tl − f.)
dt J
The active and reactive powers of the stator are given as described below [18]:
Ps = Vsd .i sd + Vsq .i sq
(9)
Q s = Vsq .i sd − Vsd .i sq
1728 M. Loucif et al.
The idea of the proposed control is a voltage oriented control (VOC), in which
the control reference is chosen to align with the q-axis voltage grid component [11,
12]:
The expressions of the derivatives stator powers (10) are easily simplified as:
Ṗs = −a1 .Ps − (ωs + a.ω)Q s + a3 .Vs .irq + b1 .Vs2 − a5 .ω.Vs .ir d − b3 .Vs .Vrq
Q̇ s = (ωs + a.ω)Ps − a1 .Q s − b3 .Vs .Vr d + a3 .Vs .ir d + a5 .ω.Vs .irq
(13)
ref ref
Where Ps and Q s are the expected active and reactive powers benchmarks.
The derivative of the equation error variables are derived as:
ė1 = Ṗsr e f − Ṗs
(15)
ė2 = Q̇ rs e f − Q̇ s
⎧
⎨ ė = Ṗ r e f + a .Ps + (ωs + a.ω)Q s − b .V 2 − a .Vs .irq + b Vs .Vrq + a .ω.Vs .i
1 s 1 1 s 3 3 5 rd
⎩ ė = Q̇ r e f − (ω + a.ω)P + a .Q − a .V .i + b .V .V − a .ω.V .i
2 s s s 1 s 5 s rd 3 s rd 5 s rq
(16)
Let us consider:
⎧ ⎧
⎨ ė01 = e1 ⎨ ė02 = e2
t
t
, (17)
⎩ e01 = λ e1 (τ )dτ ⎩ e02 = λ e2 (τ )dτ
0 0
The control algorithm proposed which is based on integral backstepping has been
introduced:
⎧
⎪ 1 re f
⎪
⎨ Vrq = b .V − Ṗs − a1 .Ps − (ωs + a.ω)Q s + b1 .Vs − a5 .ω.Vs .ird + a3 .Vs .irq − k1 .e1 − λ.e01
2
3 s
⎪
⎪ 1
⎩ Vrd = re f
− Q̇ s + (ωs + a.ω)Ps − a1 .Q s + a5 .Vs .ird + a5 .ω.Vs .irq − k2 .e2 − λ.e02
b3 .Vs
(21)
1730 M. Loucif et al.
The purpose of validating the comportment of the association of the mechanical part
with wind turbine, to extract the MPPT we use linear PI controller and electrical part
with DFIG, nonlinear integral backstepping controller is being applied to control the
active and reactive stator power. The wind turbine and DFIG parameters are provided
in the appendix.
The wind speed profile will be simulated in deterministic representation by a sum
of many harmonics is defined by the relation:
Figure 2 show the wind speed model used in this validation simulation chosen for
an average speed of 7 m/s, operate into sub-synchronous mode.
It is observed from Fig. 3, that the controller proportional and integral for
maximum power point closely tracks the generator speed as well the profile of the
wind proposed in steady state, and a slight overshoot in transient state. The operation
is always sub-synchronous mode; the slip is always positive in all profile of wind
Integral Backstepping Power Control of DFIG Based Nonlinear Modeling ... 1731
speed propose. Figure 4 the active power of the stator has a negative sign, signi-
fying that the generator produces energy and transmits it to the grid, are correctly
follows almost perfectly its reference generated by the turbine. In order to maintain
a unitary power factor on the grid side, the stator reactive power reference value will
be fixed at zero in order to ensure that the quality of the power delivered to the grid
is optimized, the power produced will track almost their reference after the end of its
transient regime. We can observe that the integral backstepping controller ensures
a exactly decoupling between them. Figure 5 the stator currents in the three phases
network generated by DFIG are sinusoidal shape and which has a excellent power
quality for the electrical grid system of frequency 50 Hz, it is very clear in the zoom
of stator currents. The rotor currents in the three phases are also sinusoidal, present
a good tracking and the frequency of the current varies with the slip shown in Fig. 6.
The electromagnetic torque variation gives the identical profile of the stator active
power.
9
Wind speed (m/s)
0
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
Time (s)
170 0.5
157
140 0.4
Generator speed (rad/s)
120
100 0.3
slip
0.2
50
*
0.1
W
W
0
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
Time (s) Time (s)
4
* x 10
4000 Ps 1
*
Qs
-2000 0
-4000
-0.5
-6000
-8000 -1
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
Time (s) Time (s)
20 10
Stator current zoom (A)
Stator current (A)
10 5
0 0
-10 -5
-20 -10
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 3 3.02 3.04 3.06 3.08 3.1 3.12 3
Time (s) Time (s)
40 20
Torque Tem (Nm)
Rotor current (A)
20 0
0 -20
-20 -40
-40 -60
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
Time (s) Time (s)
6 Conclusion
In this paper, a modeling, control and simulation of the mechanical part of a wind
turbine using a PI controller to extract the MPPT for the variable speed operating
in region two is presented. It also presents the stator active and reactive powers
control of a double-fed induction generator by considering the complete non-linear
model in the dq reference frame, using a stator voltage oriented control to simplify
the derivatives of the stator powers. Backstepping controller with integral action
applied to the rotor side converter. The slip is always positive; DFIG operates in
sub synchronous speed mode. The results of simulation are presented and discussed
confirming its feasibility of the proposed control scheme had a good follow speed
and stator powers response, regardless of variation wind speed profile.
Appendix
References
1. Abolhassani MT, Enjeti P, Toliyat M (2008) Integrated doubly fed electric alternator/active
filter (IDEA), a viable power quality solution for wind energy conversion systems. IEEE Trans
Energy Convers 23(2):642–650
2. Loucif M, Boumediene A, Mechernene A (2014) Robust nonlinear combined backstepping
sliding mode control of DFIG under wind speed variation. In: The second international
conference on electrical engineering and control applications
3. Krstic M, Kanellakopoulos I, Kokotovic PV (1995) Nonlinear and adaptive control design.
Wiley, New York
4. Krstic M, Kanellakopoulos I, Kokotovic PV (1994) Nonlinear design of adaptive controllers
for linear systems. IEEE Trans Autom Control 39:738–752
5. Yaolong T, Jie C, Hualin T, Jun H (2000) Integral backstepping control and experimental imple-
mentation for motion system. In: Proceedings of the IEEE international conference control
applications
6. Loucif M, Boumediene A, Mechernene A (2014) Maximum power point tracking based
on backstepping control of wind turbine. J Electrotehnica Electronica Automatica (EEA).
62(3):103–109
7. Ayrir W, Ourahou M, Haddi A (2018) DFIG stator reactive and active power control based
fuzzy logic. Int J Circ Syst Signal Process 12:262–267
8. El Aimani S (2004) Modélisation de différentes technologies d’éoliennes intégrées dans un
réseau de moyenne tension. Doctorat thesis, Lille, France
9. Loucif M, Boumediene A, Mechernene A (2015) Nonlinear sliding mode power control of
DFIG under wind speed variation and grid connexion. J Electrotehnica Electronica Automatica
(EEA) 63(3):103–110
10. Lihui Y, Xikui M, Dong D (2009) Hopf bifurcation in doubly fed induction generator under
vector control. Chaos Solitons Fractals 41:2741–2749
11. Bourdoulis M, Alexandridis A (2013) A new controller design and analysis of DFIG wind
turbine systems for mpp operation. In: EuroCon. Zagreb, Croatia
12. Bourdoulis M, Alexandridis A (2013) Nonlinear stability analysis of DFIG wind genera-
tors in voltage oriented control operation. In: European Control Conference (ECC), Zürich,
Switzerland
13. Loucif M (2016) Synthèse de lois de commande non-linéaires pour le contrôle d’une machine
asynchrone à double alimentation dédiée à un système aérogénérateur. Doctorat thesis,
Tlemcen, Algerie
14. Laoufi A, Hazzab A, Bousserhane IK, Rahli M (2006) Direct field oriented control using
backstepping technique for induction motor speed control. Int J Appl Eng Res 1(1):37–50
15. Haruna A, Mohamed Z, Basri MAM, Ramli L, Alhassan A (2017) 2-step integral backstepping
control of the two-rotor aero-dynamical system (TRAS). J Fund Appl Sci 9:395–407
16. El Mourabit Y, Derouich A, El Ghzizal A, El Ouanjli N, Zamzoum O (2020) Nonlinear back-
stepping control for PMSG wind turbine used on the real wind profile of the Dakhla-Morocco
city. Int Trans Electr Energ Syst 30:e12297
17. Loucif M, Boumediene A, Mechernene A (2013) Backstepping control of double fed induction
generator driven by wind turbine. In: 3rd international conference on systems and control
18. Loucif M, Boumediene A, Mechernene A (2015) Modeling and direct power control for a
DFIG under wind speed variation. In: 3rd international conference on control, engineering and
information technology
Smart Monitoring PID of Solenoid
Response Based on Fiber Squeezer
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1735
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_157
1736 S. Amrane et al.
1 Introduction
Solenoids are used in several technical applications such as robotics [1], valves
[2, 3], positioners [4], anti-braking systems [5]. They are also used in polarization
controllers to adjust the light power at the output of an optical fiber by applying
mechanical stress to it [6]. This type of actuator is characterized by a stable amplitude
of movement, which does not extend over its entire range [7]. Several studies are
devoted to extending the stability interval of this type of devices such as the nonlinear
magnetic field mapping [8], adaptive control [9], dual solenoid configuration [10],
sliding mode control [11], pulse with modulation control [12] and PID corrector that
is easy to implement [13]. However, it is useful to readjust the PID corrector in case
of the physical parameters variation of the solenoid, which affects the performance
of its response. This variation is considered to be a major problem for this type of
actuator [14–18]. The effect of fluctuations in mechanical parameters is studied in
[19].
To control the solenoid response intelligently, we propose to use an optical fiber as
a force sensor between the frame and the plunger of the solenoid. This force produces
a birefringence on the fiber, which changes the polarization state of light through the
fiber [6]. The change in the polarization state of the light results in the variation of
the light power measured by a photodiode.
First, we propose to correct the light power dynamic response at the output of the
fiber by adjusting the gains of the PID corrector. In the second step, we studied the
effect of electrical parameters variation of the solenoid the gains of the PID corrector.
The electrical parameters including resistance and inductance of solenoid. Finally,
the ANN is used to predict the values of the gains of the PID corrector to allow
maintaining the stability of the system.
The solenoid is the electromagnetic actuator that exerts pressure on the fiber. Its
structure is shown in Fig. 1
The mathematical model of the solenoid is described in detail in [20]. The mechanical
part of the system can be modeled by Eq. 1:
Smart Monitoring PID of Solenoid Response Based on Fiber Squeezer 1737
d2 x 1 dx
= Fem − k(x − x0 ) − B (1)
dt 2 m dt
Where x is the displacement of the armature in (m), m is the mass of the armature
in (Kg), Fem is the force produced by the magnetic field in (N), K is the stiffness of
the spring in (N/m), x0 is the initial air gap between the armature and the backside of
the frame in (m), and B is the system damping coefficient in (N.s/m). The electrical
part is modeled by Eq. 2:
di 1 dL(x) dx
= u − Ri − i (2)
dt L(x) dx dt
Where i(t) is the coil current in (A), L is the coil inductance in (H) that depends
on the air gap [7], and R is the coil resistance in (). Equation 3 and 4 models the
optical part.
F
δ = 6.10−5 (3)
λd
Pi
Po = (1 + cos δ) (4)
2
Where δ in (rad) is the magnitude of the phase difference between the polarized
light along the squeezing axis and it is orthogonal [21], F is squeezing force in (N),
λ is the light wavelength used in (m), d is the fiber diameter in (m), Po is the output
light intensity, and Pi is t the input light intensity in (W).
1738 S. Amrane et al.
Taking into account Eqs. 1 and 2 that respectively models the mechanical block and
the electrical block; Eqs. 3 and 4 that models the optical block, the model Simulink
is generated to simulate the solenoid as fiber squeezer (Fig. 2).
First, we use the Simulink model to generate the open-loop step response of the
simulated system. After identification from the step response, the system is looped
and corrected using a PID controller. Finally, we study the effect of fluctuation of coil
resistance and coil inductance on PID gains (Kp, Ki and Kd). The parameter variation
ranges are the resistance R for 5 to 50 and inductance L for 10 to 35 mH. The
identification of the system is done using the “tfest” function, of the Matlab/Simulink
environment, to estimate the system transfer function [22]. This transfer function Tf
has the coefficients (num, den2, den3), and can be expressed by the following form:
num
Tf = (5)
s2 + den2 s + den3
Smart Monitoring PID of Solenoid Response Based on Fiber Squeezer 1739
After estimating the transfer function of the system, the gains of the PID corrector
are determined using the “pidtune” function [23].
The Simulink model is running using the values of the solenoid with fiber squeezer
represented in the following Table 1:
Figure 3 represents the curve of variation of the light power at the output of the
fiber as a function of time when a voltage step is applied to the input of the open-loop
system (open-loop step response). The curve shows a stable response with damped
oscillations. The overshoot is relatively large (around 80%).
Table 2 The value of the PID gains used to correct the system response
Kp Ki Kd
2.9 757.5 39193
Using the pidtune function, the values of the PID gains obtained are shown in the
following Table 2:
Figure 4 represents the curve of variation of the light power at the output of the
fiber as a function of time when a voltage step is applied to the input of the closed-loop
system (close-loop step response). The curve shows a stable response. The overshoot
improved has decreased significantly (from 80% to 13%.)
The effect of the coil resistance and inductance variation the PID gains is shown in
the following maps:
Smart Monitoring PID of Solenoid Response Based on Fiber Squeezer 1741
The results of the simulation are stored in a database. Each record consists of:
– Physical parameters of solenoid: the mass m, the resistance of coil R, inductance
L, spring stiffness K, and the damping coefficient B.
– The coefficients of the transfer function: num, den2, and den3.
– The PID gains: Kp, Ki, and Kd.
The generated database is used to train the artificial neural network. The coefficients
of the transfer function represent the inputs of the neural network model while the
gains of the PID corrector represent these outputs. To predict the gains of the PID
corrector, we estimate the transfer function coefficients of the open-loop system by
using the tfest function; the parameters of this function (num, den2, and den3) are
introduced into the ANN to generate the PID gains.
4 Interpretation of Results
Figures 5, 6, and 7 represent the maps of variations of the PID corrector gains as a
function of the resistance R of the coil and the inductance L of the solenoid. These
figures show that the three gains have the same monotony when R and L vary. These
Smart Monitoring PID of Solenoid Response Based on Fiber Squeezer 1743
gains increase with R and decrease with L. We also notice that varying the coil
resistance and the coil inductance affects the solenoid response and therefore the
intensity of the output power. Consequently, it is necessary to readjust the PID gains
in the case of the resistance and inductance variation in order to maintain the response
with sufficient stability.
Using the ANN is useful to predict the gains of the PID corrector in the event of
any fluctuations in the coil resistance or the coil inductance of the solenoid. However,
it is still necessary to study and improve the performances of the ANN and its training
to minimize the error on the prediction of the PID gains.
5 Conclusion
The force applied on the optical fiber by the armature of the solenoid varies the
light power at its output. Mathematical modeling makes it possible to describe the
dynamic response of this system. PID controllers provide powerful control tools for
the transient response. The overshoot is greatly reduced from 80% to 13%. This
study also highlighted the influence of the variations of the coil resistance and coil
inductance on the solenoid response with fiber squeezer. Hence, is necessary to
readjust the PID gains in the case of the coil resistance and inductance variations.
Therefore, to keep the solenoid dynamic response in an optimal state, it suffices
to analyze its response to determine the changes that the solenoid parameters have
undergone in order to readjust the PID gains again. The use of an ANN can be used
for monitoring the PID gains to keep the solenoid dynamic response with sufficient
stability.
References
1. Takai A, et al (2013) Prototyping the flexible solenoid-coil artificial muscle, for exoskeletal
robots. In: ICCAS 2013, pp 1046–1051
2. Kajima T, Kawamura Y (1995) Development of a high-speed solenoid valve: investigation of
solenoids. IEEE Trans Ind Electron 42(1):1–8. https://doi.org/10.1109/41.345838
3. Mutschler K et al (2014) Multi physics network simulation of a solenoid dispensing valve.
Mechatronics 24(3):209–221
4. Chen M-Y et al (2010) A new design of a submicropositioner utilizing electromagnetic actuators
and flexure mechanism. IEEE Trans Ind Electron 57(1):96–106
5. Chu L, et al (2007) Study on the dynamic characteristics of pneumatic ABS solenoid valve for
commercial vehicle. In: 2007 IEEE vehicle power and propulsion conference, Arlington, TX,
USA, pp 641–644
6. Rumbaugh SH et al (1990) Polarization control for coherent fiber-optic systems using nematic
liquid crystals. J. Lightwave Technol 8(3):459–465
7. Li C et al (2016) Enhanced-performance control of an electromagnetic solenoid system using
a digital controller. IEEE Trans Contr Syst Technol 24(5):1805–1811
1744 S. Amrane et al.
8. Lim KW, et al (1994) Proportional control of a solenoid actuator. In: Proceedings of IECON
1994 - 20th annual conference of IEEE industrial electronics, Bologna, Italy, vol 3, pp 2045–
2050
9. Nataraj C, Lee D (2010) Model-based adaptive control for a solenoid-valve system. In: Volume
8: dynamic systems and control, parts A and B, Vancouver, British Columbia, Canada, pp
1009–1015
10. Yu L, Chang TN (2010) Zero vibration on–off position control of dual solenoid actuator. IEEE
Trans Ind Electron 57(7):2519–2526
11. Nguyen T et al (2007) Accurate sliding-mode control of pneumatic systems using low-cost
solenoid valves. IEEE/ASME Trans Mechatron 12(2):216–219
12. Taghizadeh M et al (2009) Modeling and identification of a solenoid valve for PWM control
applications. Compt Rendus Mécan 337(3):131–140
13. Lunge SP, Kurode SR (2013) Proportional actuator from on off solenoid valve using sliding
modes. In: Proceedings of the 1st international and 16th national conference on machines and
mechanisms, IIT Roorkee, India
14. Steiner G et al (2004) Managing uncertainties in electromagnetic design problems with robust
optimization. IEEE Trans Magn 40(2):1094–1099
15. Kabib M et al (2016) Modelling and simulation analysis of solenoid valve for spring constant
influence to dynamic response. JEAS 11:2790
16. Dülk I, Kovácsházy T (2014) Resistance estimation in solenoid actuators by considering
different resistances in the PWM paths. Period Polytech Elec Eng Comp Sci 58(3):109–120
17. Zahidi A, Said A, Azami N, Nasser N (2020) Effect of fiber and solenoid variation parameters on
the elements of a corrector PID for electromagnetic fiber squeezer based polarization controller.
IJECE 10(3):2441
18. Abedallah Z, Said A, Nawfel A, Naoual N (2019) Study and simulation of fiber press control
for solenoid parameters monitoring for Smart preventive maintenance. In: 2019 1st ICSSD,
Rabat, Morocco, pp 1–6
19. Zahidi A, Amrane S, Azami N, Nasser N (2020) Self-tuning PID of the solenoid response
based on fiber squeezer. In: 13th SITA, Rabat, Morocco, pp 1–6
20. Ikeda K, et al (2003) Endless tracking polarization controller
21. Smith AM (1980) Single-mode fibre pressure sensitivity. Electron. Lett. 16(20):773
22. Najdek K (2019) Identification of dual-active-bridge converter transfer function. Electrotech
Rev 1(3):153–156
23. Bansal H (2009) Tuning of PID controllers using simulink. Int J Math Model Simul Appl
2(3):337–344
Fuzzy Logic Based MPPT Control
of an Isolated Hybrid PV-Wind-Battery
Energy System
Abstract Recently the generations by the traditional sources are insufficient to meet
the increasing power demands due to continuous depletion of fossil energy sources,
therefore distributed generators (DGs) based renewable energy sources (RES) to
system is necessary to overcome these issues. In this study we present a stand-alone
hybrid power generation system composed by a photovoltaic and wind power genera-
tion sources with energy system storage to achieve high performance while supplying
unbalance AC load, PV system and Win turbine is operated with a DC/DC boost
converter to extract the Maximum Power Point Tracking (MPPT) using the fuzzy
logic technique. The total power generation supply the load via an inverter controlled
by pulse width modulation technique. A Simulink model of the proposed hybrid
system with the MPPT controlled Boost converter of the PV generator and Voltage
regulated voltage source inverter for standalone system is simulated in MATLAB.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1745
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_158
1746 H. Abadlia and R. Toufouti
They give promising solutions in building sustainable and ecofriendly electric power
[2]. Moreover, they decrease pollution in limiting the emission of carbon dioxide
coming from the conventional methods for energy generation based on fossil fuels.
Wind and photovoltaic power generation systems can be used for power gener-
ation as isolated system, but their benefits are considerably improved when they
are integrated into electrical grid. But the intermittent nature of these two sources
according the time of day, season and year, the random nature of these sources has
led researchers to turn to the combination of them to form the hybrid energy system
(HES) which is very interest by their reliability, their incomparable flexibility and
their attractive cost [3, 4].
HES are a good solution for alimenting rural areas located in isolated regions and
rural remote areas, where the extension of the conventional power network is more
costly and online losses become significant. But the main problem of is that (RES)
must work in a complementary way to manage the response for demand of power in
any moment.
For improvement of the complexity of the RES involved in a hybrid power system
(HPS), several researches have been devoted by several researchers in worldwide.
In order to manage the energy consumption without any problems Many studies
discussed control and energy management sources of HES presented in [5, 8, 9].
Various papers has been presented for the use sizing and optimization of hybrid
renewable systems [4, 6, 7]. For optimization HRES using intelligent techniques,
many advanced optimization techniques have been presented for the optimal design
of hybrid RES especially: Particle Swarm Optimization (PSO), Genetic algorithm
(GA) [12, 16, 23].
To extract the maximum power from RES, many MPPT algorithms are Developed
in literature such as Perturb and Observe technique (P&O), [24], incr-conductance
[25], fuzzy logic [23], sliding mode [26], also the combination of fractional order
control and incremental conductance [27]. These control strategies differ among
them in: simplicity, reject disturbances, sensors required who needed the exact
parameterization, cost and efficiency [28].
In this paper has the objective of modeling and control, inside the Matlab/Simulink
environment of isolated hybrid PV/Wind/Battery, distributed power system under
changing weather conditions to supply a variable AC load and the MPPT controller
is implemented using the fuzzy logic technique to obtained the maximum power
from the PV panels and Wind turbine to ensure the availability of power generation.
The robustness of the proposed control strategy is verified by numerical simulation.
The proposed hybrid renewables energy resources combine PV and wind energy
sources as fellows in Fig. 1:
wind generator which is used with a diode rectifier and DC/DC converter and a
PV generator includes PV panels connected to high boost converter to obtain the
Fuzzy Logic Based MPPT Control ... 1747
Fig. 1 Configuration of
proposed hybrid system
(MPPT) using fuzzy logic controller. The total energy produced from the wind and
PV sources is used to supply the AC loads through three phase inverter controlled
by pulse width modulation, the output voltage of DC/AC converter is filtered by LC
filter to given pure sine wave in the load. The IGBT invertetr is controlled by sinus
triangular modulation (SPWM).
The conception of the hybrid system is mainly dependent on the performance of each
source of energy, which must be first modeled individually and then combined to
evaluate the performance of the system.
• Modeling of PV generator. Solar energy is the most encouraging source of energy
among renewable energy, it is an attractive option due to zero CO2 emission, noise
less in operation and with lower maintenance cost [10]. The PV solar energy is
the direct transformation of light into electric power through solar cells which is
the essential element building the PV panel, PV cell is Generally a semiconductor
device composed of PN Junction almost identical to a diode [11], these cells are
connected in series to form a PV module, a number of PV modules are connected
in series to generate higher voltage which formed a string, and a number of strings
are connected in parallel to given the higher current and formed a PV array for
obtaining the required power [12].
1748 H. Abadlia and R. Toufouti
Like all the physical systems, the PV cell model is done with different levels
of exactitude depending on the users intended; this operation is realized through an
equivalent circuit with précising parameters and variables, that is why the field of PV
cell equivalent circuit model always knows advancement [13]. The one diode model
is the most used models [10, 13–15], also the model of three diode is developed
[16, 17].
We present in this paper the model of one diode due to their Simplicity [10, 18,
19], and the least number of parameters [20],this classic model consists of a current
source antiparallel with a parallel and series resistor connected in the terminal of the
diode the equivalent circuit is shown in Fig. 2 [21, 22].
Using a loop the voltage and node current laws, the generated current in the PV
generator is given:
V + I Rs V + I Rs
I = I ph − I0 ex p q −1 − (1)
nK T Rsh
Where:
Iph : The current generated by PV panel,
Io: reverse saturation diode current off the grid solar, n: the diode ideality constant.
Rs and Rsh series and shunt resistance, K: Boltzmann’s constant q: the electron
charge.
• Wind generator. Wnd turbine convert’s the mechanical energy in the wind into
electrical energy, it’s composed of turbine coupled to a generator wich transforms
the mechanical power into electrical power [29]. The generator can be a (PMSG)
or induction generator, Which are the most common generators that are used in
wind turbines [30]. Power produced by wind turbine is given as fellow: [31–33]
1
Pw = C p (λ, β)ρ AV 3 (2)
2
Where:
ρ: air density (kg/m), Cp : power coefficient, A = π × R2 : intercepting area of the
rotor blades (m), V: average wind speed (m/s), λ: tip speed ratio, β: the blade pitch
angle (in degrees).
Fuzzy Logic Based MPPT Control ... 1749
Rω
λ= (3)
V
Where:
R: radius of turbine (m), ω: angular speed (rad/s).
• Battery storage and controller by bidirectional converter. The model of battery
used in this study is been introduced in this section. The mathematical equations
based on the actual SOC of the battery as follows [4, 7, 36]:
During charge
Q Q
E batt = E 0 − K It − K I ∗ + A.exp(−bit) (5)
Q − it Q − it
During discharge
Q Q
E batt = E 0 − K It − K I ∗ + A.exp(−bit) (6)
Q − it It + 0.1Q
In this work Fuzzy logic controller is used for the control of tracking the maximum
power of the Photovoltaic generator (PVG), because of their advantages such as
robustness, simplicity to design, it require the thorough knowledge about the system
and doesn’t needs the model of the system. The proposed FLC, shown in Fig. 3 [34,
35]. In the simulation and for extract the maximum power from PV panels and Wind
turbine, A Boost Converter is widely used [36].
The inputs variables of FLC are the error E and the change of error CE at sampled
times k defined by
Where:
Pph (k) power of the PVG at sampled times k.
E(k) obvious if the load operation point is situated on the left or on the right of the
MPPT on the PV characteristic, and CE(k) give the change direction of this point.
Using Madani’s method the fuzzy inference are given in (Table 1), and the defuzzi-
fication uses the centre of gravity to compute the duty cycle D which is the output
of this FLC: [35]
n
j=1 μ D j − D j
D= n (9)
j=1 μ D j
The inputs and E(k), Ce(k) and the output D for the tracking of the maximum
power point are simulated in matlab and illustrated in Fig. 4.
Fuzzy Logic Based MPPT Control ... 1751
The proposed hybrid system has been modeled and simulated in this section, The PV
panels are modeled for 5.5 KW generated power, 6 panels are connected in series
to form string and 3 strings are connected in parallel, in total 18 panels are used for
building the solar PV system which based on the fuzzy logic MPPT algorithm. A
4 KW PMSG is driven by wind turbine, the system supplies an AC load of 8 KW
maximum considered as 6 KW principal load and additional load of 2 KW. The
simulation model permit to study different parameters of the system like voltage,
current, power also confirm the efficiency of control strategy used.
With a different metrological conditions, the simulation tests of the HRES are
implemented and interpreted. The air temperature and solar radiation have been
used as the input parameters of the PV generator. The results are obtained to evaluate
the control performance to confirm the efficiency of FLC techniques.
Figure 5 presents the variation of power generated from the PV system in different
values of solar irradiance (1000,400,800 w/m2 ) respectively, The output PV power
takes the same shape of the solar irradiance due to the direct proportional relationship
between them, the MPPT controller give maximum power output at low values of
solar irradiances.
The wind turbine generated power is depended of the wind speed variations the
output wind power depict in Fig. 6. While the wind speed is over 11 m/s, the output
power is limited to 3.5 kW. When the wind speed is less than the wind speed (6 m/s),
the output power is 1 kW.
Figure 7(a-b) the DC voltage and and load powers, we can see the Vdc voltage
I the input inverter is not affected by the power generated by the renewable sources
it’s maintained stable through MPPT controller.
The results simulation shown in Fig. 8 demonstrate that the power of hybrid system
is almost equal to the total power generated by the PV and wind generators that’s
obvious the efficiency of the fuzzy logic technique for the MPPT of PV generator.
When PHRES < Pload, the sum of power generated by the wind and the PV is
insufficient to supply the load demand. For this reason the battery start supplying
Fig. 8 Output power for PV generator, Wind generator and hybrid system
load, the state of charge start decreasing and also during discharging current become
positive.
When PHRES > Pload, there is excess power obtainable for generation. The excess
power of HRES is used for battery charging.
From these simulation results we conclude that the HRES based Fuzzy logic
MPPT controller has a good performances in the presence of irradiance and wind
speed variations, is good solution for alimenting rural areas located in isolated site
or hard-to-access locations and do not allow connection to an power network since
the investment cost.
3 Conclusion
References
1. Lopes LAC, Lienhardt AM (2003) A simplified nonlinear power source for simulating PV
panels. In: IEEE 34th annual conference on power electronics specialist, PESC 2003, vol 4.
IEEE, pp 1729–1734
2. Javed MS, Ma T, Jurasz J, Amin M (2020) Y: Solar and wind power generation systems with
pumped hydro storage: review and future perspectives. Renew Energy 148:176–192
3. Bagen, Billinton R (2005) Evaluation of different operating strategies in small stand-alone
power systems. IEEE Trans. Energy Convers 20(3):654–660
4. Ogunjuyigbe ASO, Ayodele TR, Akinola OA (2016) Optimal allocation and sizing of
PV/Wind/Splitdiesel/Battery hybrid energy system for minimizing life cycle cost, carbon emis-
sion and dump energy of remoteresidential building. Appl Energy 171:153–171. https://doi.
org/10.1016/j.apenergy.2016.03.051
5. Bakir H, Kulaksiz A (2020) A: Modelling and voltage control of the solar-wind hybrid micro-
grid with optimized STATCOM using GA and BFA. Eng Sci Technol Int J 23(3):576–584
6. Al Ghussain L, Taylan O (2019) Sizing methodology of a PV/wind hybrid system: case study
in cyprus. Environ Progr Sustain Energy 38(3):e13052
7. Diab AAZ, Sultan HM, Mohamed IS, Kuznetsov ON, Do TD (2019) Application of different
optimization algorithms for optimal sizing of PV/wind/diesel/battery storage stand-alone
hybrid microgrid. IEEE Access 7:119223–119245
8. Boussetta M, Motahhir S, El Bachtiri R, Allouhi A, Khanfara M, Chaibi Y (2019) Design and
embedded implementation of a power management controller for wind-PV-diesel microgrid
system. Int J Photoenergy. https://doi.org/10.1155/2019/8974370
9. Rezkallah M, Chandra A, Saad M, Tremblay M, Singh B, Singh S, Ibrahim H (2018) Composite
control strategy for a PV-wind-diesel based off-grid power generation system supplying unbal-
anced non-linear loads. In 2018 IEEE Industry applications society annual meeting (IAS).
IEEE, pp 1–6
10. Xiong G, Zhang J, Shi D, He Y (2018) Parameter extraction of solar photovoltaic models using
an improved whale optimization algorithm. Energy Convers Manage 174:388–405
11. Sakib S, Siddique MAB (2019) Modeling and simulation of solar photovoltaic cell for the
generation of electricity in UAE. In: 5th international conference on advances in electrical
engineering (ICAEE). IEEE, pp 66–71
12. Boukenoui R, Salhi H, Bradai R, Mellit A (2016) A new intelligent MPPT method for stand-
alone photovoltaic systems operating under fast transient variations of shading patterns. Sol
Energy 124:124–142
13. Chaibi Y, Allouhi A, Malvoni M, Salhi M, Saadani R (2019) Solar irradiance and temperature
influence on the photovoltaic cell equivalent-circuit models. Sol Energy 188:1102–1110
14. Ayang A, Wamkeue R, Ouhrouche M, Djongyang N, Salomé NE, Pombe JK, Ekemb G (2019)
Maximum likelihood parameters estimation of single-diode model of photovoltaic generator.
Renew Energy 130:111–121
15. Yahya-Khotbehsara A, Shahhoseini A (2018) A fast modeling of the double-diode model for
PV modules using combined analytical and numerical approach. Sol Energy 162:403–409
16. Khanna V, Das BK, Bisht D, Singh PK (2015) A three diode model for industrial solar cells
and estimation of solar cell parameters using PSO algorithm. Renew Energy 78:105–113
17. Allam D, Yousri DA, Eteiba MB (2016) Parameters extraction of the three diode model for
the multi-crystalline solar cell/module using Moth-Flame Optimization Algorithm. Energy
Convers Manage 123:535–548
18. Peter AG, Saha AK (2019) Electrical Characteristics improvement of photovoltaic modules
using two-diode model and its application under mismatch conditions. In: Southern African
Universities Power Engineering Conference/Robotics and Mechatronics/Pattern Recognition
Association of South Africa (SAUPEC/RobMech/PRASA). IEEE, pp 328–333
19. Bader S, Ma X, Oelmann B (2019) One-diode photovoltaic model parameters at indoor
illumination levels–a comparison. Sol Energy 180:707–716
Fuzzy Logic Based MPPT Control ... 1755
20. Gnetchejo PJ, Essiane SN, Ele P, Wamkeue R, Wapet DM, Ngoffe SP (2019) Important notes
on parameter estimation of solar photovoltaic cell. Energy Convers Manage 197:111870
21. Praiselin WJ, Edward JB (2017) Improvement of power quality with integration of solar PV
and battery storage system based micro grid operation. In: Innovations in power and advanced
computing technologies (i-PACT). IEEE, pp 1–5
22. Ayodele TR, Ogunjuyigbe ASO, Ekoh EE (2016) Evaluation of numerical algorithms used in
extracting the parameters of a single-diode photovoltaic model. Sustain Energy Technol Assess
13:51–59
23. Yilmaz U, Kircay A, Borekci S (2018) PV system fuzzy logic MPPT method and PI control
as a charge controller. Renew Sustain Energy Rev 81:994–1001
24. John R, Mohammed SS, Zachariah R (2017) Variable step size Perturb and observe MPPT
algorithm for standalone solar photovoltaic system. In: 2017 IEEE international conference on
intelligent techniques in control, optimization and signal processing, pp 1–6
25. Selvan DS (2013) Modeling and simulation of incremental conductance MPPT algorithm for
photovoltaic applications. Int J Sci Eng Technol 2(7):681–685
26. Kchaou A, Naamane A, Koubaa Y, M’sirdi N (2017) Second order sliding mode-based MPPT
control for photovoltaic applications. Sol Energy 155:758–769
27. Al-Dhaifallah M, Nassef AM, Rezk H, Nisar KS (2018) Optimal parameter design of fractional
order control based INC-MPPT for PV system. Sol Energy 159:650–664
28. Faranda R, Leva S (2008) Energy comparison of MPPT techniques for PV Systems. WSEAS
Trans Power Syst 3(6):446–455
29. Chow JH, Sanchez-Gasca JJ (2020) Power System Modeling, Computation, and Control. Wiley,
Hoboken
30. Arturo Soriano L, Yu W, Rubio JDJ (2013) Modeling and control of wind turbine. Math Probl
Eng. https://doi.org/10.1155/2013/982597.
31. Patel MR (2005) Wind and solar power systems: design, analysis, and operation. CRC Press
32. Salehi J, Badr AA (2019) Dynamic modeling and optimal control of a wind turbine with doubly
fed induction generator using imperialist competitive and artificial bee colony algorithms. J
Circ Syst Comput 28(04):1950070
33. Natsheh EM, Albarbar A, Yazdani J (2011) Modeling and control for smart grid integration
of solar/wind energy conversion system. In: 2011 2nd IEEE PES International conference and
exhibition on innovative smart grid technologies, pp 1–8
34. Takun P, Kaitwanidvilai S, Jettanasen C (2012) Maximum power point tracking using fuzzy
logic control for photovoltaic systems. In: World congress on engineering 2012, 4–6 July 2012,
London, UK, vol 2189. International Association of Engineers, pp 986–990
35. Cheikh MA, Larbes C, Kebir GT, Zerguerras A (2007) Maximum power point tracking using
a fuzzy logic control scheme. Rev Energies Renouvelables 10(3):387–395
36. Feddaoui O, Toufouti R, Jamel L, Meziane S (2020) Fuzzy logic control of hybrid systems
including renewable energy in microgrids. Int J Electr Comput Eng (IJECE), 10(6). (2088-8708)
Implementation of a Digital Control
for PV Power Systems
Badreddine Lahfaoui
1 Introduction
Solar panels are the main sources of any photovoltaic system. They are connected in
parallel or in series to remove their operating voltage or current. The energy supplied
by the system can be used to charge batteries which will provide electricity when
needed.
It can also be used by directly connecting the modules to the load without batteries
(i.e. for a solar pump, water is used as storage), or by plugging them into an electrical
network. It is also possible to combine the output of the PV field with other energy
sources such as a generator or a wind turbine which will serve as back-up, if there is
not enough sunlight.
Photovoltaic systems can be either (Autonomous system without battery,
Autonomous system with battery, Hybrid PV/generator system) the stand-alone
systems, not connected to a power grid. Or the other three types (PV system on
B. Lahfaoui (B)
Mohammed First University Oujda, EST Oujda, LGEM, 60000, Oujda, Morocco
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1757
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_159
1758 B. Lahfaoui
This subset converts solar energy into electrical energy. It is made up of the following
elements:
– A 30 W “Monocrystalline” type photovoltaic solar module.
– An aluminum structure tilting from 15 to 75°.
– 3 safety connectors (4mm) (+, −, PE).
– A non-return diode located in the module connection box.
The ETM53930 solar panel consists of 39 monocrystalline silicon cells connected
in series. Table 1 presents the electrical specifications, of this module, which are
obtained under standard test conditions (STC):
– Solar irradiation of 1000 W/m2 .
– 1.5 air masses.
– And temperature of 25° C.
We also present the electrical performance of this solar module illustrated by me
in the labs which represents on the one hand the electrical power as a function of
voltage and on the other hand the electric current as a function of voltage.
Implementation of a Digital Control for PV Power Systems 1759
Table 1. PV system
Parameters Values
characteristics
Weight 3.5 kg
Dimensions 633 × 427 × 35 mm
Maximum power voltage (Vmp) 19.4V
Maximum power current (Imp) 1.55A
Open circuit voltage (Voc) 23.8V
Short circuit current (Isc) 1.74A
Temp.coeff. of Isc (Tk Isc) 0.06%/°C
Temp.coeff. of Voc (Tk Voc) −0.397%/°C
Temp.coeff. of Pmax (Tk Pmax) −0.549%/°C
Normal operating cell temperature 44.4+− 2 °C
Wanting to exploit the optimum power of the solar module ET-M53930.For this
we have implemented the numerical control algorithm in the DSPACE 1104 card.
Figure 2 represents the experimental hardware formed by:
– PV source 30W monocrystalline solar module - ET-M53930
– Artificial solar source (W/m2 )
– One Boost
– Load
– Current sensor
– Voltage sensor
– DSPACE 1104 card
– Computer with ControlDesk software
The Perturb and Observe Numerical control has been implemented on a DSPACE
1104 card to perform power regulation.
Implementation of a Digital Control for PV Power Systems 1761
The algorithm of this method as illustrated in the Fig. 3 begins with the measure-
ment of the two parameters current and voltage Vpv (k), Ipv (k) to calculate the instan-
taneous power value at the instant k and the previous power at the instant k-1. Also
the algorithm calculates the difference in voltages Vpv = Vpv (k) − Vpv (k−1).
Then the command checks the sign of the value of Ppv and following the sign
verification of the value of Vpv the algorithm increments or decrements the duty
cycle Dpv with a value of Dpv .
This algorithm is experimentally translated into a programming language using
the Simulink tool from MATLAB software and is then sent to the DSPACE control
card handled via the Control Desk software.
1762 B. Lahfaoui
Table 3 Measuring
Equipment Parameters
equipment
Current sensor 10A to 1V and 100A to 1V
Voltage sensor ST1000
We present above the experimental results of the various study tests in the absence
of disturbance in Figs. 4 which include the power at the terminals of the solar panel
Ppv as function as the value of the duty cycle Dpv .
The optimum power is reached for all of these tests through the use and
implementation of the MPPT control in the DSPACE 1104 card.
– The maximum power at the out of the solar panel oscillates approximately around
18W.
– The optimal duty cycle oscillates approximately around 0.7.
The experimental results illustrated in Fig. 4 are captured by Controdesk software
over a time interval of 500 s and a solar irradiance equal to approximately 1000W/m.
Implementation of a Digital Control for PV Power Systems 1763
Fig. 4 Tracking the maximum power using DSPACE 1104 Card (In Controldesk).
1764 B. Lahfaoui
5 Conclusions
The maximum optimization of renewable energy sources has become essential for
socio-economic and environmental reasons. For this, various research works are
initiated on several sources such as photovoltaic sources, wind sources, and other
sources.
Our research work described in this paper presents the optimization of a photo-
voltaic system driven by a numerical control to satisfy a resistive load under solar
irradiance and constant temperature.
The experience was carried out in the laboratory, LGEM, using several control
measuring equipment and artificial sources.
A validation of the experimental results shown in this paper proves the success of
our proposed architecture.
References
1 Introduction
The use of classical loads as reinforcement of polyamides is more and more inad-
visable thanks to their environmental, economic and sustainable development draw-
backs [1, 2]. Scientific research is more and more interested in a new category of
materials called “eco composites” or “bio composites” meeting the requirements
of sustainable development, manufacturers and environmental constraints. They
are based on natural bio fillers also offering a lower cost and density with a high
biodegradability.
In this work, the Argan Nut Shell (ANS) particles will be used in order to give them
the good value in the socioeconomic and environmental context. The Nylon PA66
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1765
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_160
1766 A. Moumen et al.
Table 1 Characteristics of
ANS Values
ANS
Content 20 Wt%
Orientation 0,45 and 90°
Aspect ratios 0.5,2 and 20
used from the family of thermoplastics has a light weight and good thermomechanical
properties [3–5]. Several Nylons can be quoted: 6,11,12,46 and 66 etc. [6–11]. They
are more used in transportation, automotive, textile and packaging industries. The
PA66 is one of the most used Nylons. Its synthesis is made by the hexamethylene
diamine and adipic poly condensation.
Filled Nylon has better characteristics in terms of strength and stiffness. We will
study in this work the optimal properties of the Nylon PA66 bio filled by ANS
Particles with the Mori Tanaka and finite element methods by varying the content,
orientation and aspect ratios of particles.
2.1 Materials
The ANS particles are rich in cellulose, hemi cellulose, lignin and Ashes by (25.7%,
34.3%, 34.5% and 5.4% respectively) as determined by Essabir et al. [2]. The Table 1
presents the properties of the particles used.
The Nylon 66 used is from the family of thermoplastics having a light weight and
good thermo-mechanical properties [5, 12, 13].
2.2 Methods
We will use this model to predict the mechanical behavior of Nylon 66 bio filled by
ANS particles. It is advantageous compared to other homogenization models such
as Voigt, Tsai Pagano and Reuss etc. [14–16] it gives more efficiency and reliability
of predicted behaviors. The homogenization parameters are cited in the Table 2.
In order to have better convergence, we are going to use a minimal, maximal and
final times respectively of 0.01, 0.1 and 1 s. The values chosen in Table 2 are the
result of several numerical simulations leading to good convergence and reliability
of the results.
Towards a Digital Modeling of the Optimal Mechanical Properties … 1767
Fig. 1 Geometry of Nylon PA66 bio filled by ANS particles a Cylinder particles, b Ellipsoid
particles, c Sphero cylinder particles
performances is due to the potentially large set of constraint equations that has to be
imposed.
A macroscopic uniaxial strain state in the 1-direction is imposed (macroscopic
uniaxial stress state) as presented in the Fig. 3.
C. Incrementation and Mesh
We are going to use a max of 20 increments with initial, minimum, maximum and
final times respectively of 0.1, 0.01, 0.1 and 1. The mesh used is the voxel non-
conforming, having several advantages in terms of convergence, flexibility of use
and the best accuracy [20].
Towards a Digital Modeling of the Optimal Mechanical Properties … 1769
3.2 Density
The density of ANS particles is around 1.29 g/cm3 which is lower compared to
mineral or inorganic fillers such as 2.6g/cm3 clay [2] and talc 2.8 g/cm3 [2] showing
a lightness of our bio load which can benefit the polyamides studied.
The Table 3 and the Fig. 5 show the density of the eco-composite based on ANS
particles compared to other composites loaded with mineral and inorganic particles
(clay and talc).
Fig. 4 Mesh of Nylon PA66 bio filled by ANS particles a Cylinder particles, b Ellipsoid particles,
c Sphero cylinder particles
In order to study the effect of ANS particles orientation on the behavior of the polymer
studied, we modeled different orientations ranging from 0° to 90°.The orientation
tensor is defined by:
⎡ ⎤
a11 a12 a13
ai j = ⎣ a12 a22 a23 ⎦ (1)
a13 a23 a33
In order to obtain a trace equal to 1, we define a trace tolerance of 0.1. If the result
is far from [1-tol; 1 + tol], the numerical modeling corrects the orientation tensor by
this way (Table 4):
⎡ ⎤ ⎡ a ⎤
a11 a12 a13
11
trace(a)
a12 a13
⎢ ⎥
ai j = ⎣ a12 a22 a23 ⎦is corrected to a = ⎣ a12
a22
trace(a)
a23 ⎦ (3)
a33
a13 a23 a33 a13 a23 trace(a)
⎡ ⎤
0.3016 −0.0157 −0.0218
ai j (cylindrical) = ⎣ 0 0.3919 −0.0138 ⎦ (5)
0 0 0.3065
⎡ ⎤
0.3016 −0.0113 −0.0477
ai j (ellipsoid) = ⎣ 0 0.3263 0.00797 ⎦ (6)
0 0 0.3721
⎡ ⎤
0.3472 0.00351 0.00244
ai j (sphero cyl) = ⎣ 0 0.3517 0.02548 ⎦ (7)
0 0 0.3012
If we want to express the aligned Representative Volume element, it’s better to choose
the fixed orientation tensor to save computing time. We define the angles θ and φ.
The first one is closed to 90° and the second one varies from 0° to 180° depending
on the orientation in the plane (1–2).
The following figures obtained by the finite element method present the elastic
behaviors of composites for 0°, 45 and 90° orientations (Fig. 6).
The analysis of the curves in the figure below shows a slight stiffness in terms
of Young modulus and the stress at break for cylindrical, sphero cylindrical and
ellipsoid particles at 0°. This rigidity decreases slightly away from 0°.The less rigid
behavior is the same for cylindrical and sphero-cylindrical particles at 45° and 90°
whereas for ellipsoid particles, the behavior at 45° is slightly higher than 90°.
In order to validate the results obtained for the orientation effect, a comparison
between the results of theMori-Tanakaand finite element methods is carried out as
presented by the following figures:
The analysis of the curves of the Figs. 7, 8 and 9 obtained numerically shows that
the results obtained by the finite element method are similar to those obtained by
the Mori-Tanaka model. The rigidity decreases slightly by moving away from 0° in
terms of Young modulus and the stress at break for cylindrical, sphero-cylindrical
and ellipsoid particles.
Compared to the orientation, it is slightly advised to use cylindrical and sphero
cylindrical morphologies especially when using particles in orientations close to 90°.
1772 A. Moumen et al.
Fig. 6 Particle orientation effects for the three morphologies using the finite element method
Fig. 7 Elastic curves for cylindrical and sphero cylindrical particles at 0° by the FE method and
M-T model
In order to study the effect of ANS particles aspect ratio on the behavior of the
polymer studied, we modeled several aspect ratios (0.5, 2 and 20). The Figs. 10
and 11 present respectively the tensile curves obtained by theMori-Tanaka and finite
element methods for these different aspect ratios.
Towards a Digital Modeling of the Optimal Mechanical Properties … 1773
Fig. 8 Elastic curves for cylindrical and sphero cylindrical particles at 45° and 90° by the FE
method and M-T model
Fig. 9 Elastic curves for ellipsoid particles at 0°, 450 and 90° by the FE method and M-T model
The results illustrated in Figs. 10 and 11 shows that the Mori-Tanaka and finite
element methods give the same results for aspect ratios of 0.5,2 and 20.The two
numerical methods clearly show that the rigidity of the bio composite improves by
increasing the aspect ratio of Argan Nut Shell. At an aspect ratio of 20, the rigidity is
much better. The stress at break is greater with these high aspect ratios. These results
are valid for all the morphologies studied (cylinder, sphero cylinder and ellipsoid).
1774 A. Moumen et al.
4 Conclusion
In the present work, the importance of bio loading was studied in order to have green
eco composites meeting the expectations of researchers in terms of respecting the
environment, regulations and sustainable development requirements.
Towards a Digital Modeling of the Optimal Mechanical Properties … 1775
The Mori Tanaka and finite element methods were used to characterize the
mechanical behavior of Nylon 66 bio filled by ANS particles. These two digital
methods have proved a great ability to predict the behavior of the eco composite
studied. They are reliable, simple in use and have more speed in computing.
At 20 Wt% of particles and by increasing their aspect ratios, the mechanical
properties of the eco composites studied are optimum. By moving away from 0°, the
rigidity slightly reduces. The use of these bio fillers will replace the use of traditional
fillers and move towards environment friendly eco composites.
Future Scope
In this study, we determined the mechanical behavior of Nylon PA66 bio loaded by
ANS particles. It will be interesting to study the thermal behavior and apply other
tests on this eco composite such as fatigue and aging tests. Other bio loads of animal
or plant origin can be studied for the reinforcement of this Nylon using experimental
tests and digital modeling.
References
12. Leconte N, Bourel B, Lauro F, Badulescu C, Markiewicz E (2020) Strength and failure of an
aluminum/PA66 self-piercing riveted assembly at low and moderate loading rates: experiments
and modeling. Int J Impact Eng 142:103587
13. Moumen A, Lakhdar A, Mansouri K (2020) Numerical study of the mechanical behavior of
polyamide 66 reinforced by argan nut shell particles with the finite element method and the
mori-tanaka model. Int J Adv Trends Comput Sci Eng 9:7723–7730. https://doi.org/10.30534/
ijatcse/2020/115952020
14. Moumen A, Jammoukh M, Zahiri L, Mansouri K (2020) Study of the optimal micromechanical
behavior of a polymer reinforced by snail shell particles using the mori-tanaka numerical model.
In: 2020 IEEE international conference of Moroccan geomatics (Morgeo). IEEE, pp 1–6
15. Lakhdar A, Moumen A, Zahiri L, Jammoukh M, Mansouri K (2020) Experimental and numer-
ical study of the mechanical behavior of bio-loaded PVC subjected to aging. Adv Sci Technol
Eng Syst J 5:607–612
16. Chihaoui B, Serra-Parareda F, Tarrés Q, Espinach FX, Boufi S, Delgado-Aguilar M (2020)
Effect of the fiber treatment on the stiffness of date palm fiber reinforced PP composites:
macro and micromechanical evaluation of the young’s modulus. Polymers (Basel) 12:1–21.
https://doi.org/10.3390/POLYM12081693
17. Moumen A, Jammoukh M, Zahiri L, Mansouri K (2020) Numerical modeling of the thermo
mechanical behavior of a polymer reinforced by horn fibers. Int J Adv Trends Comput Sci Eng
9:6541–6548. https://doi.org/10.30534/ijatcse/2020/342942020
18. Lakhdar A, Jammoukh M, Zahiri L, Mansouri K, Moumen A, Salhi B (2020) Numerical and
experimental study of the behavior of PVC material subjected to aging. In: 2020 1st international
conference on innovative research in applied science, engineering and technology (IRASET).
IEEE, pp 1–6
19. Yu G, Shi B, Shen Y, Gu J (2020) A novel finite element method for predicting the tensile
properties of 3D braided composites. Mater Res Exp. 6:125626
20. Rao MVP, Raj D, Harursampath D, Renji K (2020) Estimation of material properties of metal
matrix composites using finite element method in the presence of micromechanics damages.
Mater Today Proc 21:1135–1143
21. Sokolov A, Schetinin V, Kozlov M (2020) Surface finite element for imperfect interface
modeling in elastic properties homogenization. Key Eng Mater 833:101–106
Experimental Evaluation of MEMS
Accelerometers Integrated
into Smartphones: A Case Study
of Bearing Condition Monitoring
Abstract Over the last decade, smartphones have become an essential part of
numerous applications in daily life activities. Contrary to professional equipment,
smartphones are available to everyone and are not expensive, which is why it is a
valuable opportunity to exploit smartphone capabilities for professional applications.
This paper presents an experimental evaluation of integrated MEMS accelerometers
into smartphones to assess their performance for vibration monitoring, with the
expectation to use them for educational and modest industrial purposes. This study
presents an evaluation of two different devices having different specifications, which
evaluates the extracted features of the vibration data collected from both devices.
A shaking table test was carried out to compare the two MEMS accelerometers
results with professional equipment. Finally, a case study was conducted to monitor
the condition of two bearings running at low rotational speed (600 rpm) using the
vibration data acquired through both smartphones. Experimental results indicate that
smartphones can be a useful instrument for monitoring the condition of bearings.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1777
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_161
1778 A. Ait Ben Ahmed et al.
1 Introduction
Vibration analysis is very important to identify and to monitor the health of mechan-
ical systems [1], but many people consider that the cost of vibration analysis equip-
ment is very expensive and not at hand for all. Contrary to professional equipment
(Pro. Equip.), the smartphones are characterized by low cost compared with profes-
sional equipment, and by their several functionalities. So, it’s beneficial to look
for replacing professional equipment with smartphones to use them for educational
purposes such as mechanical vibration practical works in the university curriculum,
or modest industrial purposes such as monitoring the rotating machinery operating
under low speed, and correcting the unbalance defect in rotors [2, 3].
Each day, smartphones have more and more features to add to existing devices.
The vast majority of smartphones contain a wide variety of sensors, it has become a
favored tool in the world and even an important part of daily life activities [4]. The
sensor system of smartphones contains motion sensors such as accelerometers and
gyroscopes, magnetic field measurements, light detection, and proximity detection.
Besides, sound can be detected using the microphone when the images are captured
using the camera. Sensors have been used in smartphones since their introduction,
the new smartphones are powerful devices that are always on hand and have access
to a wireless connection at anytime and anywhere [5].
Nowadays, most smartphones are programmable and equipped with autonomous,
inexpensive, small, and low energy sensors, this gives smartphones more powerful
benefits. This paper aims to presents an experimental evaluation of the accelerometer
integrated into the smartphones, to assess its performance in vibration analysis, and
compare them with professional equipment. This study shows the evaluation of two
accelerometers with different specifications and compares their performance with
professional equipment, the comparison will be based on the time and the frequency
features extracted from the vibration data recorded by each accelerometer.
Pre-processing the accelerometer data is very important to ensure the accuracy and
the quality of the measured data. To measure the real movement of the device it is
recommended to remove the gravity component of the acceleration measurements.
This can be achieved by applying a high-pass filter to eliminate the effect of gravity
from three directions [7]. For noise filtering, a Butterworth Infinite Impulse Response
(IIR) bandpass filter has been applied to remove noise from the vibration signal, with
appropriate cut-off frequencies according to the range of phenomena to be observed.
In this section, we will present the shake table to evaluate the smartphone’s
accelerometer and compare its results with those of professional equipment.
1780 A. Ait Ben Ahmed et al.
Figure 2 illustrates the experimental setup which contains the rotor system that
produces a uniform vibration (≈16 Hz) using an unbalanced mass. The vibration
of the rotor causes the vibration of the table. The smartphone and professional
equipment capture the vibrations of the table along the Z-axis direction of the
smartphone.
Figure 3 shows the vibration signal obtained using the smartphone and using the
professional equipment during the evaluation test in the Z-axis direction.
Smartphone
1 Pro. Equi.
0.8
0.6
Acceleration (m/s²)
0.4
0.2
-0.2
-0.4
-0.6
-0.8
-1
0 0.05 0.1 0.15 0.2 0.25 0.3
Time (s)
Fig. 3 Smartphone vs. professional equipment vibration signal during the evaluation test
Experimental Evaluation of MEMS Accelerometers … 1781
Table 3 Time-domain
Features New Sp. Old Sp. Pro. Equip.
statistical features of
vibration data (Fastest mode, Mean −0.0004 0.0027 −0.0258
Z-axis direction) RMS 0.5585 0.5437 0.4937
Kurtosis 1.5260 1.6093 1.5134
Skewness 0.0024 −0.0082 0.0167
Variance 0.3125 0.2962 0.2433
Crest factor 1.6090 1.8119 1.4382
Statistical features of the time-domain, such as mean, root mean square (RMS),
Crest factor, and variance were generally used to identify distinctions between the
vibration signals. More advanced statistical features, such as skewness, and kurtosis
are also applied as features [9, 10], all these features are used in this study to evaluate
the smartphone vibration data. Also, the frequency spectrum of the vibration data
generated by the Fast Fourier transform (FFT) algorithm is exploited for extracting
and evaluating other features of the vibration data [11].
Tables 3 and 4 show the time-domain features of the vibration signal recorded by
the old and new smartphone compared with the professional equipment results. The
1782 A. Ait Ben Ahmed et al.
Table 4 Time-domain
Features New Sp. Old Sp. Pro. Equip.
statistical features of
vibration data (Game mode, Mean 0.0007 −0.0003 −0.0258
Z-axis direction) RMS 0.5765 0.5144 0.4937
Kurtosis 1.6153 1.7024 1.5134
Skewness −0.0551 −0.1139 0.0167
Variance 0.3330 0.2649 0.2433
Crest factor 1.7442 1.8042 1.4382
measurement concerned just the two first modes (Fastest and Game, see Table 2)
because the sampling rate of each mode is higher than the frequency of vibration
phenomena (Shake table frequency ≈16 Hz).
As shown in Tables 3 and 4, the statistical features of the vibration data,
RMS, Kurtosis, Variance, and Crest factor are more stable for both devices and in
both sampling frequency modes, compared to professional equipment results. And
concerning the frequency-domain, the new smartphone has a clear spectrum than the
spectrum obtained by the old smartphone, which is very noisy. So, this difference
was produced owing to the resolution and the accuracy of the two sensors.
The resolution and the sampling rate of the smartphone’s accelerometer play an
important role in the accuracy of the measurement and the amount of noise in the
vibration signal. Also, the type of sensor mounting on the structure affect the obtained
results, it’s recommended to be adhesive mounting.
Figure 4 shows the frequency-domain of the vibration signal recorded by the old
and new smartphone for the Fastest and Game modes in the Z-axis direction.
Based on Fig. 4, the new smartphone has a clearer spectrum than the spectrum
obtained by the old smartphone which is very noisy. It can therefore be said that the
resolution is responsible for this difference. Besides, the sampling frequency plays
an important role in the spectrum clarity, so that the spectrum of the Fastest mode
(100 Hz) is better than the spectrum of the Game mode (50 Hz).
Generally, the smartphone presents acceptable results, but the low sampling rate
of the smartphone’s accelerometer (Max 100 Hz) still the main disadvantage which
limits the interval of using smartphones in the professional areas. So, in this case, it
is recommended to evaluate the vibration phenomena in which the main frequency
is located below half of the Smartphone sampling rate (i.e. 50 Hz).
In this section, two bearings were examined to identify the degree of damage, the
first was a healthy bearing and the second was defective with a localized fault in
one ball of the bearing, as shown in Fig. 5. Bearing fault signals are considered as
amplitude modulated signals. Where the faults excite the resonance frequencies of
the bearing structure. For this reason, it is difficult to identify fault frequencies in the
Experimental Evaluation of MEMS Accelerometers … 1783
0.9 0.9
(a) (b)
0.8 0.8
0.7 0.7
Amplitude (m/s²)
0.6 0.6
Amplitude (m/s²)
0.5 0.5
0.4 0.4
0.3 0.3
0.2 0.2
0.1 0.1
0 0
0 5 10 15 20 25 30 35 40 45 50 0 5 10 15 20 25
Frequency (Hz) Frequency (Hz)
0.6 0.6
(c) (d)
0.5 0.5
Amplitude (m/s²)
0.4 0.4
Amplitude (m/s²)
0.3 0.3
0.2 0.2
0.1 0.1
0 0
0 5 10 15 20 25 30 35 40 45 50 0 5 10 15 20 25
Frequency (Hz) Frequency (Hz)
Fig. 4 Frequency-domain of vibration signal recorded by both smartphones, a New Sp. Fastest
Mode, b New Sp. Game Mode, c Old Sp. Fastest Mode, (d). Old Sp. Game Mode
Fig. 5 The rotor system with the smartphone mounted tightly to the bearing
signal spectrum. Also, envelope analysis is not useful in our case due to the lower
sampling frequency (Fs) of the smartphone accelerometer. Therefore, we confine to
perform a statistical evaluation.
The smartphone was mounted near the bearing to acquire vibration data in the
three directions (X, Y, and Z axes). Time-domain features, selected in Sect. 4, will
be extracted from the vibration data to examine the health of the two bearings where
the rotor is operating at 600 rpm (10 Hz).
1784 A. Ait Ben Ahmed et al.
0.1 0.8
(a) (b)
0.08
0.6
0.06
0.4
0.04
Acceleration (m/s 2)
Acceleration (m/s 2)
0.2
0.02
0 0
-0.02
-0.2
-0.04
-0.4
-0.06
-0.6
-0.08
-0.1 -0.8
0 5 10 15 0 5 10 15
Time (s) Time (s)
Fig. 6 Time-domain of smartphone vibration data measured in the vertical direction (Y-axis),
a Healthy bearing, b Faulty bearing
Table 5 Time-domain statistical features of vibration data (Old Sp., X-, Y-, and Z-axis)
Features Direction Healthy bearing Faulty bearing Ratio (Healthy/Faulty)
RMS X-axis 0.0479 0.0980 48.8%
Y-axis 0.0120 0.0956 12.5%
Z-axis 0.020 0.0447 44.7%
Kurtosis X-axis 3.8118 5.5676 68.4%
Y-axis 3.1493 15.299 20.5%
Z-axis 3.4848 5.6071 62.1%
Variance X-axis 0.0023 0.0096 23.9%
Y-axis 0.0001 0.0091 1.09%
Z-axis 0.0004 0.002 20.0%
Crest factor X-axis 3.5314 4.7288 74.6%
Y-axis 3.2845 7.7377 42.4%
Z-axis 2.6394 5.6562 46.6%
and the dispersion has also increased around the reference mean value, and the signal
becomes more impulsive.
In sum, for efficient monitoring, it is recommended to make a schedule to systemat-
ically acquire vibration data at predefined points and to establish graphs of the evolu-
tion as a function of time for each feature in order to monitor the bearing degradation
and to create a smartphone-based system for bearing condition monitoring.
6 Conclusion
References
3. Ait Ben Ahmed A, Touache A, El Hakimi A, Chamat A (2020) A new hybrid method for rigid
and flexible rotor balancing without phase response measurements. Aust J Mech Eng. https://
doi.org/10.1080/14484846.2020.1842616
4. Feng M, Fukuda Y, Mizuta M, Ozer E (2015) Citizen sensors for SHM: Use of accelerometer
data from smartphones. Sensors 15(2):2980–2998. https://doi.org/10.3390/s150202980
5. Cerutti S, Magenes G, Bonato P (2010) Special section on smart wearable devices for human
health and protection. Institute of Electrical and Electronics Engineers. https://doi.org/10.1109/
TITB.2010.2048937
6. Google: Sensor—Android Developers. https://developer.android.com/guide/topics/sensors/
sensors_overview. Accessed 29 Nov 2020
7. Google: Sensor—Android Developers. https://developer.android.com/guide/topics/sensors/
sensors_motion. Accessed 11 Sept 2020
8. SensorDelay Enum (Android.Hardware)—Microsoft Docs. https://docs.microsoft.com/en-
us/dotnet/api/android.hardware.sensordelay?view=xamarin-android-sdk-9. Accessed 09 Dec
2020
9. Soong TT (2004) Fundamentals of probability and statistics for engineers. Wiley, Hoboken
10. Caesarendra W, Tjahjowidodo T (2017) A review of feature extraction methods in vibration-
based condition monitoring and its application for degradation trend estimation of low-speed
slew bearing. Machines 5(4):21
11. Chu E, George A (1999) Inside the FFT black box: serial and parallel fast fourier transform
algorithms. CRC Press, Boco Raton
MPPT Using Adaptive Genetic-Fuzzy
Logic Control for Wind Power System
Abstract This study proposes a fuzzy-based MPPT controller used to extract the
maximum power point (MPPT) from the WECS using doubly fed induction generator
(DFIG). The MPPT was accomplished by applying control to the generator speed
and blade pitch angle. The parameters of the FLC are optimized by using a genetic
algorithm (GA). Initial simulation results prove that the proposed MPPT algorithm
possesses good dynamic and steady state performances under different operating
conditions.
Keywords Wind turbine · Doubly fed induction generator (DFIG) · Fuzzy Logic
Controller (FLC) Genetic algorithm optimization
1 Introduction
Several challenges are still confronting the large-scale adoption of WECS (wind
energy conversion system). The most urgent challenge concerns the wind speed
variations at the various spots over the blades covered area. Thus, making direct
sensing of wind speed a challenging task [2, 9]. In addition, deploying mechanical
sensors rises the overall costs of WECS, while reducing its reliability and roughness
[2–4]. In order to meet these challenges, the author of [1] have proposed an estimation
of the actual wind speed indirectly by using electric indicators, such as output electric
power or generator shaft speed using a tachometer which are reasonably easy to
measure according to [4]. The authors of [5–7] combined these results with a dynamic
model of wind turbine. However, these dynamic equations are highly nonlinear and
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1787
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_162
1788 C. Alaoui et al.
Grid
Fig. 1 Wind turbine, doubly fed induction generator and power electronic converters block diagram
having hard to find parameters. For that reason, the wind speed is usually estimated
by using some kind system identification, data mining, state observer and/or some
advanced fuzzy techniques [1, 3].
Modern wind turbine systems include embedded controllers to predict the best
pitch angle for the wind turbine blades to operate under various wind speed scenarios.
This will serve to adjust the pitch angle. The objective of the presented project is
to implement a controller that will drive the pitch angle of the blades of the wind
turbine system to the preset pitch angle in order to regulate the input wind power.
Hence extracting the maximum allowable power without damaging the wind turbine
nor the electric generator.
In this project, we consider a wind power system comprising of a wind turbine,
a doubly fed induction generator (DFIG) connected to a three-phase grid through
appropriate AC-DC and DC-AC converters [8]. The proposed system is illustrated
in Fig. 1.
The wind turbine transforms the kinetic energy of the air particles, captured by the
blades, into mechanical power that rotates the electric generator. The static converters
transmit the electric power to the grid. This is accomplished by the drive train system,
rotor/stator of the induction machine and an AC-DC converter coupled with a DC-
AC converter through a capacitor. The control system reads the output power and
adjusts the pitch angle of the blade system of the wind turbine in order to regulate
the output power.
MPPT Using Adaptive Genetic-Fuzzy Logic Control … 1789
The power accessible from the wind in a wind turbine system is expressed by (1).
P = 0.5ρ Aυ 3 (1)
where ρ is air density (1.225 kg/m3 ), A is the area swept by the blades (m2 ), and v
is wind speed (m/s). Figure 2 shows the wind power as a function of wind speed for
an uncontrolled system.
The power available in the air molecules cannot be completely transferred to a
mechanical power in the shaft of the turbine; it is rather limited by the Betz threshold
of 59%. This limit is expressed by the power coefficient C p of the wind turbine.
It is a function of the blade pitch angle (β) and the blade tip speed ratio (λ). The
mechanical power available at the shaft of the wind turbine becomes (2).
where β is the blade pitch angle, C p is the power coefficient of the turbine, λ is the
blade tip to speed ratio (TSR). C p is a nonlinear function of β and λ and given in (3)
and (4).
116 −21
C p = 0.5176 − 0.4β − 5 e λi + 0.0068λ (3)
λi
1 1 0.035
= − (4)
λi λ + 0.08β 3β + 1
Cp
β
λ
Fig. 3 The power coefficient Cp -TSR curve for different pitch angle values
Finally, blade tip to speed ratio-TSR is defined in (5), where ω is the blade angular
velocity (rad/s) and R is the wind turbine’s blade radius (m).
ω t R
λ= (5)
υ
Figure 3 illustrates the nonlinear relationship between the blade tip speed ratio
(TSR) and the power coefficient (C p ), with the pitch angle (β).
Blade pitch regulation means to adjust the angle of attack of the blades of a wind
turbine rotor in order to control the electric power produced by the DFIG. The
main objective of the blade pitch control system is to keep the rotor speed within
MPPT Using Adaptive Genetic-Fuzzy Logic Control … 1791
Fig. 5 Output power vs. turbine rotational speed for various pith angle values
Fig. 6 Tip to speed ratio based MPPT for wind turbine block diagram
1792 C. Alaoui et al.
(λopt ) of the turbine. The system would then be able to extract the maximum attain-
able power as shown in Fig. 6. The optimum rotational speed is determined as ωopt =
λopt V w /R.
Several control approaches [13, 14] have been used that use the Fuzzy Logic Control
(FLC) for WECS MPPT applications, as illustrated in Fig. 7. The foremost advantage
of such controllers is that their respective parameters can be modified rapidly in
response to sudden variations in the system dynamics. When the weather conditions
changes, the performance of a FLC based MPPT is robust. However, their robustness
depends heavily on the knowledge base (KB) and the experience of the operator
in choosing the rule base (RB), the membership functions and tolerable levels of
output errors. Additionally, the memory requisite poses additional limitations in
FLC application. This is the main reason to include a proper optimization algorithm,
such as genetic algorithm (GA), in order to optimize the fuzzy engine of the FLC,
and hence the adaptive FLC.
The adaptive fuzzy controller system has many advantages when used for non-linear
system. Some of these advantages are:
• Automatically adapts it parameters to new values as the environmental settings
are highly dynamic.
• The parameters are altered to get the desired and fixed output values.
MPPT Using Adaptive Genetic-Fuzzy Logic Control … 1793
Fig. 8 Altering the scaling function. Having the scaling function of 200, the fuzzy subset positive
P = [0 0.5 0.7], however, with the scaling function of 400, P = [0 0.25 0.35]
Fig. 9 Adapting the fuzzy set definition. The fuzzy subset P = [0 0.5 0.7] becomes P = [0 0.19
0.4]
the wind power system, represented by the ‘controlled system’ block, is evaluated,
and its output power is used as fitness value for the genetic algorithm (GA). The
GA block produces the optimized parameters for the fuzzy controller block that, in
combination with the state parameters, produces the PWM signal to the control the
converters. Fuzzy modeling of type Takagi–Sugeno was selected in this study since
accuracy was a preferred criterion for the MPPT system.
MPPT Using Adaptive Genetic-Fuzzy Logic Control … 1795
The simulation was executed on a wind power system with specifications listed
in Table 1. The proposed genetic-fuzzy control system for MPPT was done by
Matlab/Simulink. The simulation results, shown in Fig. 11, compare the theoret-
ical values of the generator power and the shaft speed, with the actual output power
and rotational speed as obtained by the proposed method. The wind speed was set
to change from 6 to 8 m/s at t = 2.5 s and from 8 to 10 m/s at t = 4.5 s during the
Matlab simulation.
The Simulation results, as illustrated in Fig. 11, reveal that the proposed method
has good dynamic response, with steady state performances during the course of the
MPPT process.
Fig. 11 Theoretical vs. simulated output power de the wind power system
4 Conclusion
In this project, a fuzzy logic controller (PLC) was proposed to wind turbine pitch
control in order obtain the MPPT throughout the wind speed values. A genetic-fuzzy
control-based for DFIG wind power system has been discussed and modeled. The
efficacy of the projected MPPT strategy has been proved by initial Matlab/Simulink
simulations.
References
1. Zhiqiang X, Qinghua H, Ehsani M (2012) Estimation of effective wind speed for fixed speed
wind turbines based on frequency domain data fusion. Sustain Energy IEEE Trans 3(1):57e64
2. Tan K, Islam S (2004) Optimum control strategies in energy conversion of PMSG wind turbine
system without mechanical sensors. Energy Convers IEEE Trans 19(2):392e9
3. Wei Q, Wei Z, Aller JM, Harley RG (2008) Wind speed estimation based sensorless output
maximization control for a wind turbine driving a DFIG. Power Electron IEEE Trans
23(3):1156e69
4. Meyer DG, Srinivasan S, Semrau G (2013) Dynamic wind estimation based control for small
wind turbines. Renew Energy 50:259e67
5. Kodama N, Matsuzaka T, Tuchiya K, Arinaga S (1999) Power variation control of a wind
generator by using feed forward control. Renew Energy 16(1):847e50
6. Van der Hooft E, Van Engelen T. Estimated wind speed feed forward control for wind turbine
operation optimisation. In: Conference estimated wind speed feed forward control for wind
turbine operation optimisation
7. Boukhezzar B, Siguerdidjane H (2009) Nonlinear control with wind estimation of a DFIG vari-
able speed wind turbine for power capture optimization. Energy Convers Manag 50(4):885e92
8. SimPowerSystems for Use with Simulink (2004) User’s Guide Version 4, MathWorks, Inc.
Natick, MA
MPPT Using Adaptive Genetic-Fuzzy Logic Control … 1797
9. Buehring IK, Freris LL (1981) Control policies for wind-energy conversion systems. IEE Proc
C Gener Transm Distrib 128:253
10. Abdullah MA, Yatim AHM, Tan CW, Saidur R (2012) A review of maximum power point
tracking algorithms for wind energy systems. Renew Sustain Energy Rev 16(5):3220–3227
11. Knight AM, Peters GE (2005) Simple wind energy controller for an expanded operating range.
IEEE Trans Energy Convers 20:459–466
12. De Kooning JD, Gevaert L, Vyver JVD, Vandoorn TL, Vandevelde L (2013) Online estimation
of the power coefficient versus tip-speed ratio curve of wind turbines. In: 39th annual conference
IEEE industrial electronics society, pp 1792–1797
13. Simoes MG, Bose BK, Spiegel RJ (1997) Fuzzy logic based intelligent control of a variable
speed cage machine wind generation system. IEEE Trans Power Electron 12:87–95
14. Hilloowala RM, Sharaf AM (1996) A rule-based fuzzy logic controller for a PWM inverter in
a stand-alone wind energy conversion scheme. IEEE Trans Ind Appl 32:57–65
Development of an Automated
Measurement Platform for an Electrical
Traction System Test Bench
Abstract In this paper, a test bench for electrical traction systems is presented
and discussed. A simulation model in Matlab/Simulink is proposed to bypass the
limitations imposed by the mentioned test bench. The developed Simulink model
allows the emulation of the electric vehicle performance for urban driving conditions
using ECE 15 drive cycle. An automated measurement platform is made, with the
aim of improving the test bench, by making it possible to reproduce the simulation
results with fidelity and ease.
1 Introduction
Today’s global society is very concerned with reducing the negative effects of road
transportation on the environment due to toxic emissions and greenhouse gases.
Therefore, these types of vehicle emissions are legally regulated in multiple countries
across the world, in order to comply with the stricter regulations that are anticipated
to be adopted in the future, vehicle manufacturers are obliged to invest in various
fuel-saving technologies. This has resulted in an increased interest in the electrifi-
cation of vehicles, primarily hybrid electric vehicles (HEVs) which can reduce fuel
consumption compared to conventional vehicles, but also battery electric vehicles
(BEVs) [1, 2]. The latter offer high driveline efficiency and do not emit exhaust
gases, which is why they have been considered CO2 free in regulations to date [3, 4].
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1799
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_163
1800 A. El Houre et al.
Electric vehicles are expected to replace regular vehicles in the near future [5]. It is
therefore very important to conduct research and train a new generation of engineers
in this field.
There are different methods for developing and studying electric vehicles, each
with its own advantages and disadvantages. One of the best-known methods is the
test bench based study. This approach allows researchers to evaluate the perfor-
mance of electric propulsion and develop new control strategies in closed scientific
research environments [6, 7]. In order to maximize the benefit of the test bench,
automation of test measurement and data’s graphical representation is necessary
since it reduces time and effort compared to manual testing, in addition to providing
increased productivity and reliability and increased test predictability [8].
In this paper an experimental test bench of an electric traction motor is presented,
then a Matlab/Simulink test bench simulation model is made and discussed in order
to emulate the performance of the electric vehicle’s traction motor for on-road driving
conditions. Afterwards, a platform for the automation of measurement and graphical
display of data generated from the electric traction motor test bench is built.
Figure 1 illustrates the experimental test bench concerned in this study, and Fig. 2
shows the structure of the experimental test bench. The test bench consists of a
Development of an Automated Measurement Platform … 1801
3 Simulation Model
Figure 3 shows the synoptic diagram of the drive cycle based test that allows the
emulation of the vehicle performance on city driving conditions, and Fig. 4 demon-
strates the Simulink model created to reproduce the mentioned test. The simulation
idea of the test bench is based on Fajri’s work [6] with certain modifications to suit
1802 A. El Houre et al.
our case. The motor and dynamometer model parameters used in the simulation were
extracted experimentally in order to get closer to reality.
The drive cycle model generates speed and acceleration references corresponding
to the ECE 15 cycle that represents city driving conditions [12]. These references are
sent to the vehicle’s dynamic model, which describes all the forces exerted on the
vehicle, and calculates the speed and the torque references for the speed controlled
motor and the torque controlled dynamometer, respectively. A mechanical shaft
model is used to simulate the mechanical coupling of the two machines. The test
bench model generates the torque and speed data of the motor; this data will be used
to plot the torque/speed curve. Figure 5 illustrates the simulation results.
Figure 5(a) shows that the dynamometer motor follows almost perfectly the torque
reference, meaning that the torque regulation is working as intended. Figure 5(b)
indicates that the motor torque opposes the dynamometer motor torque with the
same amplitude. However, the torque peaks were caused by the sudden change of
the resistive torque, which affects the speed profile as well. Figure 5(c) confirms the
speed regulation of the motor since it’s almost the same as the reference.
In order to verify the validity of the current simulation model results, another
simulation method was used, which is in our case the Advanced Vehicle Simulator
(ADVISOR) [13].
A vehicle with almost the same specifications as our vehicle is designed and
simulated using ADVISOR. This simulated vehicle follows perfectly the speed profile
of the ECE 15 drive cycle.
Figure 6 indicates that the drive motor torque resulted from the simulation model
and ADVISOR on the ECE 15 drive cycle are nearly similar most of the time.
However, since the vehicle’s specifications for both simulations are not exactly the
same because of the limitations imposed by ADVISOR and the fact that our test bench
model doesn’t include regenerative braking as opposed to ADVISOR, the results of
the two simulations were therefore slightly distorted, particularly with regard to the
negative torques.
Development of an Automated Measurement Platform … 1803
Fig. 5 Simulation results for the drive cycle based test. a Dynamometer motor reference and
generated torque. b Drive motor torque. c Drive motor reference and generated rotational speed
Fig. 6 Drive motor torque comparison between our Simulink model and ADVISOR
1804 A. El Houre et al.
Fig. 7 Simulation result of the drive motor torque/speed data points corresponding to the ECE 15
drive cycle
These results show that our simulation model emulates the electric-drive vehicle
performance on ECE15 drive cycle correctly.
After verifying our simulation results using ADVISOR, the torque/speed data
points of the drive motor according to the ECE 15 cycle were plotted as shown in
Fig. 7.
This simulation model proved that the simulated drive motor based on the actual
motor parameters is able to achieve the whole ECE 15 operating points accurately.
So as to verify this result experimentally, the resistive load and the speed of the drive
motor should be manipulated in a way to reproduce the same operating points as
Fig. 7.
Another way for doing so would be to reproduce the experimental drive motor
acceleration performance and projecting it on the ECE 15 operating points, and if
these points fall within the acceleration torque/speed envelope, then we could say
that these ECE 15 operating points are achievable by the drive motor experimentally.
In both cases, the automation of measurement and graphical display of data gener-
ated from the test bench, would be a priority since we want to be able to follow the
performance of the test bench faithfully and with ease.
The main purpose of the measurement circuit is measuring the required electrical
signals and conditioning them to be usable by the DAQ device.
The most important quantities in our test bench are: current, voltage, speed, torque,
power and efficiency. For the torque T, it’s estimated using the measured load current
and dynamometer torque constant, as described by Eq. (1) which was validated
experimentally. Concerning the speed N, it’s calculated based on the frequency of
the hall effect signal of either the motor or the dynamometer [14, 15], and the pole
number of one of the two machines, as determined by the relation (2).
T = kt × I (1)
Where kt is the torque constant in N.m/A of the dynamometer motor, and I is the
measured load current.
120
N= × f (2)
p
f is the frequency of the hall effect signals, and p is the pole number of the
machine.
As for the power and efficiency, they are deducted from the voltages, currents,
speed and torque.
The calculations of the quantities stated above are implemented in the LabVIEW
interface.
The control and configuration section of the interface Fig. 9, allows the user to
control the execution of the program, to insert the necessary quantities to calculate
the torque and the speed.
It is possible, via this sheet, to save the test results in an Excel file and to display the
elapsed time of the test and errors that might occur during its execution. Figure 9(b)
indicates the XY representation of speed versus torque, source current vs torque,
mechanical power versus torque, efficiency versus torque, and finally the efficiency
map of the electrical drive. The dashboard section showed in Fig. 9(b) visualizes the
important quantities extracted from the test bench in real-time.
Fig. 9 The LabVIEW interface. a The control and configuration section of the interface. b The XY
data representation section. c The dashboard section
Development of an Automated Measurement Platform … 1807
5 Conclusion
In this paper, an experimental test bench was introduced and discussed. This test
bench presents a drawback in terms of the emulation of the electric vehicle perfor-
mance for on-road driving conditions. To overcome this, a simulation model was
proposed in the Matlab/Simulink environment and validated using ADVISOR. This
simulation contains a model of the test bench based on real experimental parameters
of the drive motor, dynamometer, and vehicle dynamics, which allows the repro-
duction of the drive cycle based test. The next objective would be reproducing the
data operating points resulted from this simulation using the actual test bench to
validate that the electrical drive is suitable for our application. In order to do so, an
automated measurement platform was developed to monitor and to follow faithfully
and effectively the performance of the test bench.
Acknowledgement This work was funded by ALTRAN Maroc, a Global Engineering and R&D
Services Company, and was supported by the Moroccan Research Institute for Solar Energy and
New Energies (IRESEN).
References
12. A reference book of driving cycles for use in the measurement of road vehicle emissions:
version 3. IHS (2009)
13. Wipke K, Cuddy M, Burch S (1999) ADVISOR 2.1: a user-friendly advanced powertrain
simulation using a combined backward/forward approach. IEEE Trans Veh Technol 48:1751–
1761
14. Naveen V, Isha TB (2017) A low cost speed estimation technique for closed loop control
of BLDC motor drive. In: 2017 international conference on circuit power and computing
technologies (ICCPCT), pp 1–5
15. Dumitru D. Module 8: Speed Control - NXP Community, https://community.nxp.com/t5/NXP-
Model-Based-Design-Tools/Module-8-Speed-Control/m-p/727476. Accessed 29 Oct 2020
Simulation of a Guidance Law
on a ROS-Based Nonholonomic Mobile
Robot
1 Introduction
Wheeled mobile robots play a major role in both industrial and service robotics,
particularly when autonomous motion capabilities are required, however developing
a fully autonomous robot that can navigate in unknown environments is problem-
atic due to difficulties that span dynamics modeling, on-board perception, trajectory
generation, and optimal control, not to mention the heavy computational bill that may
results from such methods. Although several studies denounce limitations that are
imposed by the structural obstruction of the kinematic model, for instance designing
feedback laws, little attention has been drawn to address these obstacles. This paper
develops a waypoint trajectory tracking method that demonstrates the possibility to
overcome the limitation imposed by the kinematic model. One of its most attractive
features is computational simplicity, and the ability to control the accuracy of gener-
ated paths by changing the number of waypoints. in the last decade, Mobile robot
path planning has drawn lot of interest from researchers in deep learning, however,
as reported by [1] real-time processing require a large neural network with multiple
parameters, which represent a computational burden, even for modern CPU, in [2]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1809
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_164
1810 N. Rezoug and M. Zerikat
they used fuzzy PID controller for autonomous motion of a differential drive robot,
in order to solve the path and time optimization problem they developed in [3] a four-
layer Neural network as a controller, in [4] they used as an experimental platform a
car like mobile robot where they focused on designing an effective neural network
controller. By using linearized model predictive control (LMPC) in [5] they dealt
with the system physical constraints, therefore, overcoming the limitations caused
by nonholonomic constraints of mobile robot. In [6] time varying state feedback
control law based on back-stepping technic was implemented on the Turtlebot to test
the performance of the studied strategy in [7] they consider the problem of robust
adaptive trajectory tracking control for a mobile robot dealing with system uncer-
tainties. In order to achieve robust tracking two control strategies were used in [8] a
proportional integral control and sliding mode control, in order to simplify the struc-
ture of the control scheme, a new kinematic control approach was used for position
tracking. Path following is basically dealing with the problem of designing a guidance
law that steers the robot to follow a desired path, based on the approach presented
in [9], and with the sole purpose of avoiding any latency related problem that may
occur from computational complexity, we aim in this paper to design guidance law,
in order to simplify the structure of the control system.
The reminder of this paper is organized as follows: in Sect. 2 a thorough description
of the experimental test bed is presented, Sect. 3 presents the model of the robot and
outline the problem that motivated this work, Sect. 4 describe the strategy used
to control the robot, Sect. 5 shows the simulation results and comparison, Sect. 6
contains the conclusion and future work.
2 Experimental Testbed
Typically, the guidance law allows an accurate path tracking in two-dimension frame.
In this section, we briefly review the kinematic model of a differential drive mobile
robot, and present the concept of waypoint guidance method which provides a
rigorous formulation of the problem of steering the differential drive mobile robot
along a desired path.
ẋr = Va cos(ψ),
ẏr = Va sin(ψ), (1)
ψ̇r = ω;
Where xr andyr denote the robot position in the fixed frame, and ψ it’s orientation.
Va is the linear velocity, and ω is the angular velocity.
ψe = ψ p − ψ (4)
By differentiating (2) with respect to time and using (1), it follows that
ψ̇e = ω (6)
In the following we illustrate the functional structure of the control system design
(Fig. 4).
The control law is given by
ẋ = f (x(t), u(t))
The input u is the angular speed ω, our objective is to design a control law u(t)
that stabilizes the system and consequently steers the Turtlebot to the reference path.
T
For the equilibrium point xeq = deq , 0 , u eq = 0, where deq is arbitrary, we
obtain a linearized state space model of the system as follows
x = A p x + bpu (8)
Where
0 −Va 0
Ap = , bp = (9)
0 0 −1
The regulated output y is the cross-track error d, i.e., the output vector is
c= 10 (10)
e I = yg − y (11)
yg denotes the reference output. Since the system output is the distance from the
robot to the waypoint reference position, naturally we set yg = 0 in order for the
robot heading direction to converge toward the target location.
e I = 0 − cx (12)
Simulation of a Guidance Law on a ROS-Based Nonholonomic Mobile … 1815
Ẋ = AX + Bu
ẋ A p 02×1 x b
= + u (13)
ė I −c 0 eI 0
+∞
J = ∫ X T (t)Q X (t) + Ru 2 (t) dt (14)
0
Q(3 × 3) and R(1 × 1) are the state weighting matrix and the control weighting
matrix, respectively. As shown in [12], using the state feedback control law (15) will
result in the minimization of the cost function (14)
A T P + P A − P B R −1 B T P + Q = 0 (16)
5 Experimental Results
To show the effectiveness and accurate tracking capabilities of the Turtlebot using
the guidance law presented in this paper, we used two kind of reference trajectories
generated in a 2D frame, square trajectory and circular trajectory. The simulations
were carried out on the Turtlebot Kobuki under Gazebo with parameters as follows.
Diameter: 351.5 mm; Height: 124.8 mm; Weight: 2.35 kg.
The simulation was conducted under GAZEBO to assess the effectiveness of the
trajectory tracking controller, results of the simulation are shown in Figs. 5 and 8.
The environment GAZEBO was used for simulating a virtual “Turtlebot” seeking
to track a squared and circular trajectory. Test results are discussed in the following.
1816 N. Rezoug and M. Zerikat
The robot initially positioned at the origin (0, 0), is first required to reach the
closest point on the desired trajectory and then follow the given waypoints.
The square trajectory is located at (2, 2); (4, 2); (4, 4); (2, 4). Figure 5 shows
that the robot tracks the given waypoints with a negligible error around the edges of
the square trajectory, the error is mainly due to the imposed range tolerance, which
basically define the exact reach the robot has to get within in order to move to the
next waypoint.
Next, we show the results as presented in [13], where the robot seems to give a
satisfactory result under MATLAB environment, as shown in Fig. 6, however the
simulation result under gazebo world does not follow the desired trajectory Fig. 7.
The results output in [13] can be explained by the fact that under gazebo world
various dynamic properties are taken into account, while under MATLAB simulation
many of these dynamics are not considered. Which justifies the efficiency of the
adopted control law in MATLAB simulation, and the poor performance under Gazebo
world.
Simulation of a Guidance Law on a ROS-Based Nonholonomic Mobile … 1817
Fig. 6 Circular trajectory diagram of the Turtlebot under MATLAB as shown in [13]
Fig. 7 Circular trajectory diagram of the Turtlebot under Gazebo as shown in [13]
The circle is centered in (2, 2), we can see in Fig. 8 that the robot started tracking
with an initial error, which is natural due to the large initial heading error, so basically
there was a time of transition in which the tracking error of heading direction was
stabilized.
1818 N. Rezoug and M. Zerikat
6 Conclusion
In this paper, a waypoint guidance law was developed for differential-drive wheeled
mobile robots. An accurate tracking is obtained from various reference trajectories,
3D simulation was developed in Gazebo world using MATLAB-Simulink for moni-
toring and data processing. In future work, we plan to test the proposed scheme in
a real robot, adding more challenging tasks, for instance the difficulty to identify
external dynamics, which will result in a limited dynamic compensation, hence the
necessity for an additional control scheme to compensate the dynamic properties
in order to have more accurate tracking performance. We will aim to achieve the
goal of composite nonlinear control and guarantee that the output tracking error will
ultimately converge to zero. In order to further reduce the oscillation amplitude and
frequency, friction compensation schemes can be incorporated, an extra control level
will focus on improving the robot’s behavior such as obstacle avoidance. To sum up,
this paper lay the ground work for future projects with the sole purpose of improving
the performance of ground robots.
Simulation of a Guidance Law on a ROS-Based Nonholonomic Mobile … 1819
References
1. Vanhoucke V, Senior A, Mao MZ (2011) Improving the speed of neural networks on CPUs.
In: Deep learning and unsupervised feature learning, NIPS
2. Heikkinen J, Minav T, Stotckaia AD (2017) Self-tuning parameter fuzzy PID controller for
autonomous differential drive mobile robot. In: 2017 XX IEEE international conference on
soft computing and measurements (SCM), pp 382–385. https://doi.org/10.1109/SCM.2017.
7970592
3. Parhi DR, Singh MK (2009) Real-time navigational control of mobile robots using an artificial
neural network. Proc Inst Mech Eng Part C: J Mech Eng Sci 223(7):1713–1725
4. Dewi T, Risma P, Oktarina Y, Roseno MT (2017) Neural network controller design for a mobile
robot navigation; a case study. In: 2017 4th international conference on electrical engineering,
computer science and informatics (EECSI)
5. Fetter Lages, W, Augusto Vasconcelos Alves J (2006) Real-time control of a mobile robot using
linearized model predictive control. In: Proceedings Volumes, vol 39, no 16, pp 968–973. ISSN
1474-6670, ISBN 9783902661173
6. Besseghieur, K.L., Tr˛ebiński, R., Kaczmarek, W., Panasiuk, J.: Trajectory tracking control for
a nonholonomic mobile robot under ROS. In: Journal of physics: conference series, volume
1016, 6th international conference on mechatronics and control engineering (ICMCE 2017)
7. Xin L, Wang Q, She J, Li Y (2016) Robust adaptive tracking control of wheeled mobile robot.
Robot Auton Syst 78:36–48 ISSN 0921-8890
8. Becerra HM, Colunga J, Romero J (2018) Simultaneous convergence of position and orientation
of wheeled mobile robots using trajectory planning and robust controllers. Int J Adv Rob Syst
15:172988141875457. https://doi.org/10.1177/1729881418754574
9. Souanef T (2019) Adaptive guidance and control of small unmanned aerial vehicles. https://
doi.org/10.13140/RG.2.2.21153.56166
10. Kobuki Documentation: relase 2.0, Dabit industries
11. Corke P (2015) Integrating ROS and MATLAB [ROS Topics]. IEEE Robot Autom Mag 22:18–
20
12. Zak, SH (2003) Systems and control. vol 198. Oxford University Press, Oxford
13. Bensaci C, Zenir Y, Pomorski D (2017) Control of mobile robot navigation under the
virtual world matlab-gazebo. In: A special issue of the international conference on advanced
Engineering in Petrochemical Industry (ICAEPI’17), Skikda Algeria, vol 2, no 4
Speed Sensorless Direct Torque Control
of Doubly Fed Induction Motor Using
Model Reference Adaptive System
Abstract This study emphasis on a direct torque control (DTC) strategy with a speed
estimator utilizing the Model Reference Adaptive System (MRAS) for a doubly fed
induction machine (DFIM). The suggested adaptive reference model uses the values
obtained as the disparity between two voltage-to-current ratios along the d and q axes
for a hypothetical resistance quantity. The resulting sensorless formulation is totally
independent of any expression of stator/rotor resistance. And thereby, DTC command
brings numerous benefits over traditional control strategies like field-oriented control
(FOC) due to its clear architecture, quick dynamic response and less system param-
eter dependency. The suggested control algorithm is simulated and evaluated in
MATLAB/Simulink for a small speed range program.
1 Introduction
The DFIM is one of the most common high-power devices, particularly for the fields
of wind and hydropower generation [1]. It is also used in systems of attraction, marine
drives and electric and hybrid cars [2]. But the difficulty of these calculations and
the combination of torque and flux cause many difficulty.
In order to solve these issues, some researchers are focused on designing rigorous
control strategies. Among the first methods, the field orientation control technique
offers a decoupling of flux and torque. This covers a diverse range of modula-
tion velocity, quick torque reactions and high efficiency for several variations in
load [3]. Despite the appealing features of the FOC control, its execution involves
several coordinate transformations and a reliable configuration of a variety of PI
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1821
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2_165
1822 M. E. Mahfoud et al.
controllers, and requires an external pulse width modulator (PWM) [1]. To solve
the problems of FOC control, TAKAHASHI implemented Direct Torque Control
in the early 1980s [4]. In comparison to the FOC, the DTC approach focused on
hysteresis comparators and turn tables provides many advantages such as easy execu-
tion, limited reliance on parameters, without any need for current controllers and
several coordinate transformations [5].
While in most industrial uses, speed and/or position detection is needed for robust
and significant command. As a consequence, DFIM speed regulation using the DTC
uses a speed sensor for its application, this speed sensor [6]:
• Diminishes machine performance;
• Raises the cost of the control device and is challenging to install
• Diminished reliability;
• Extra area for assembly,
• The wire.
That is why speed estimation methods are an amazing option. Literature has
proposed various methods such as Luenberger observer, Kalman filter, slipping mode
control, Model Reference Adaptive Systems (MRAS) and artificial intelligence tech-
niques [6, 7]. Of all the techniques, the MRAS Observer for Sensorless DFIM drives
is well known and most desirable due to its flexibility, simple physical perception
and tolerance to variations in system parameters [8].
The MRAS methodology therefore involves many forms of methods based on the
parameter selected for the estimate. Among these methods, the R-MRAS estimator
uses a resistance value corresponding to the voltage/current ratio that provides high
independence and high efficiency [9]. The goal of this paper is to combine the benefits
of DTC control with the efficiency of the MRAS approach for speed control and
estimation.
This paper is organised as follows: Sect. 2 introduces a DFIM modelling. Section 3
addresses traditional DTC approach based on switching tables and hysteresis
controllers. Section 4 addresses speed and flux measurement requirements. Section 5
discusses and interprets simulation curves using MATLAB/SIMULINK environ-
ment. Finally, Sect. 6 summarises the conclusions and recommendations for future
work.
The DFIM model for DTC is the two-phase system represented in (α, β) coordinates.
The complexity of (a, b, c) representation of the motor is decreased by the Concordia
transformation [5]. Figure 1 indicates the spatial structure of these various frames of
reference.
The angle θ r represents the position of the rotor in relation to the reference axis
of the stator, while the angles θ and θ s show the relative positions of the direct axis
respectively with respect to the rotor and stator axes.
Speed Sensorless Direct Torque Control … 1823
β d (Direct axis)
(Quadrature axis) ω
q a (Rotor)
ω
θ
θ
θ α 0 A
O (Stator)
With:
θs = θr + θ (1)
dθs dθr dθ
= + ⇒ ωs = ωr + ω (2)
dt dt dt
The DFIM stator and rotor voltage can be represented in (α, β) reference frame
[10]:
⎡ ⎤ ⎡ ⎤⎡ ⎤ ⎡ ⎤ ⎡ ⎤
vsα Rs 0 0 0 i sα ψsα 0
⎢v ⎥ ⎢ ⎥ ⎢i ⎥ d ⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ Rs 0 0
sβ 0 ⎥ ⎢ sβ ⎥ ⎢ ψsβ ⎥ dθ ⎢ 0 ⎥
⎢ ⎥=⎢ ⎥.⎢ ⎥ + .⎢ ⎥+ .⎢ ⎥ (3)
⎣ vr α ⎦ ⎣ 0 0 Rr 0 ⎦ ⎣ ir α ⎦ dt ⎣ ψr α ⎦ dt ⎣ −ψrβ ⎦
vrβ 0 0 0 Rr irβ ψrβ ψr α
The expression of stator and rotor flux in diphase reference frame (α, β):
⎡ ⎤ ⎡ ⎤ ⎡ i sα ⎤
ψsα Ls 0 M. cos θ −M. sin θ
⎢ψ ⎥ ⎢ ⎢i ⎥
⎢ sβ ⎥ ⎢ 0 Ls M. sin θ M. cos θ ⎥⎥.⎢ sβ ⎥
⎢ ⎥= ⎢ ⎥ (4)
⎣ ψr α ⎦ ⎣ M. cos θ M. sin θ Lr 0 ⎦ ⎣ ir α ⎦
ψrβ −M. sin θ M. cos θ 0 Lr irβ
d
With Tem = TL + J. + f. (6)
dt
1824 M. E. Mahfoud et al.
Takahashi proposes the DTC technique for electric motors to address field-oriented
control problems and particularly the sensitivity to variation in machine parameters.
DTC has a lot of advantages, such as reduced dependence on outside disruptions and
machine parameters, simple operation and a relatively quicker response of dynamic
torque to other controls [5]. The power of flux and torque in the stationary structure (α,
β) is decoupled. The switching table is used to choose a suitable voltage vector [10].
The choice of switching states is specifically linked to flux and torque variations.
The electromagnetic torque and flux amplitudes in hysteresis regulator bands are
thus reduced. These regulators control these two dimensions separately. Hysteresis
controller inputs are flux and torque errors and their performance defines the neces-
sary voltage vector to strike the inverter at each switching cycle. The inverter supplies
eight voltage vectors, two of which are null and void. These vectors are chosen from
a switching table that was calculated with the flux, torque and flux vector positions
[11]. The distribution of these vectors is seen in Fig. 2 (α, β).
The flux error εψ is introduced to a two-levels hysteresis comparator, which
produces the binary output (Hψ = 0,1) describing the ideal flux progression. The
torque error εTem is also inserted into a three-levels hysteresis controller, which
produces the variable (HTem = −1, 0, 1) at its output, which corresponds to the
path of the torque creation required.
The switching table provided in Table 1 is drawn up by the variables HTem and
Hψ division as well as the sector presenting details on flux vector locations.
The MRAS Theory is the most popular method used to observe machine parameters
and states using only the measurements of stator and rotor voltage and current [7].
The MRAS observer system consists of two systems with separate mechanisms for
estimating the same state variable depending on the separate inputs [12]:
• The first technique is focused on the equations in the stationary frame (α, β), being
independent of the speed to be estimated; It’s the reference model.
• the adaptive model is the second structure, which relies on the value to be
estimated, is described in the (d,q) reference (Fig. 3).
The values extracted from the models are compared and the variance is used to
operate an adjustment process generating the value to be computed. Centered on
this error, the adjustment mechanism generates the speed value calculated by the PI
controllerIf the algorithm is constructed right, the steady state would go to 0. [6].
Figure 4 reflects.
MRAS observer design.
• Reference model:
vαr vβr
R∗ = − (7)
i αr i βr
• Adaptive model:
vdr vqr
R= ∗ − ∗ (8)
i dr i qr
1826 M. E. Mahfoud et al.
Reference
Model
Comparator PI
α,β
d,q Adaptive R
Model
Integrator
In DTC, depending on the errors in the control variables, the switching tables indi-
cate the positions of the inverters switches to generate the three-phase voltages that
attack the DFIM. The discrepancy between approximate variables and their refer-
ences measures these errors. Therefore, flux and torque accuracy is also very critical
to ensure satisfactory performance. Many variables must be defined, the stator and
the rotor current calculated, while the rotor and stator voltages may depend on the
states (Sa , Sb and Sc ).
The electromagnetic torque, rotor flux and stator flux are estimated by following
equations [5]:
T em = p.(ψ sα .i sβ − ψ sβ .i sα ) (9)
ψ̂r = ψ̂r2α + ψ̂rβ
2
(10)
ψ̂s = 2 + ψ̂ 2
ψ̂sα sβ (11)
ψ̂rβ
θr = ar ctg( ) (12)
ψ̂r α
ψ̂sβ
θs = ar ctg( ) (13)
ψ̂sα
The block diagram for the calculation of flux and torque is illustrated in Fig. 4.
Speed Sensorless Direct Torque Control … 1827
Rotor flux
estimation
Stator flux
estimation
Torque T
estimation
According to these results, the rotational speed follows the reference value
perfectly and quickly, It attains the reference value after tr = 0.28 s without over-
shooting and static errors at the start. On the other hand, the electromagnetic torque
leads with very rapid dynamics the set reference. At runtime, the torque is maximum
and maintained in constant state at nearly value 0. Also, Where a load torque appli-
cation at t = 1.8 s, a speed change of 1.8 rad/s occurs the time to reject this disruption
is quite quickly equivalent to 30 ms, which illustrates the effectiveness of this regu-
lation technique against changes in load torque. But, the electromagnetic torque has
ripples of the range of Tem = 2.7 N.m. Furthermore, the evolution of each flux
represents a circular course with a constant radius identical to the reference.
Speed Sensorless Direct Torque Control … 1829
The stator and rotor current aspects are sinusoidal in shape and react well to
the changes caused by the load torque. It can also be remembered that the currents
frequencies are equivalent to the autopilot equation’s rotational speed. Finally, the
positions flux are perfectly respected.
6 Conclusion
This paper has presented a speed sensorless DTC scheme for DFIM powered by two
voltage inverters. The goal is to boost traditional DTC efficiency. The hierarchical
DFIM model and the DTC strategy theory were introduced. The control technique
and the speed and flux calculation are comprehensive. After device controls have
been checked with MATLAB/SIMULINK the major results are as below.:
• The MRAS approach can be applied successfully in DFIM drives.
• The fast response, torque dynamics and robustness of the classic DTC are
maintained.
The future studies will concentrate on experimental DTC-MRAS control valida-
tion utilising the dSPACE DS1104 platform.
References
© The Editor(s) (if applicable) and The Author(s), under exclusive license 1831
to Springer Nature Switzerland AG 2021
S. Motahhir and B. Bossoufi (eds.), Digital Technologies and Applications,
Lecture Notes in Networks and Systems 211,
https://doi.org/10.1007/978-3-030-73882-2
1832 Author Index
U Z
ul Huda, Anwaar, 925 Zahidi, Abdallah, 1735
Zahra, Syeda Anam, 925
V Zamzoum, Othmane, 1235
Varghese, Levin, 271 Zaoui, Mohamed, 233
Venna, Deepa, 939 Zellou, Ahmed, 335
Vijaykumar, Chaya, 813 Zerikat, Mokhtar, 1809
Zerraf, Soufiane, 111
W Ziane, Abderrezzaq, 1147
Wadie, Bendali, 1317 Ziani, El Mostafa, 1445
Wahabi, Aicha, 447 Zine-Dine, Khalid, 279
Wakrim, Layla, 1013 Zulfiqar, Ayesha, 469