Springer
Springer
Introduction
to Artificial
Intelligence
Imaging Informatics for
Healthcare Professionals
Series Editors
Peter M. A. van Ooijen, University Medical Center Groningen,
University of Groningen, GRONINGEN, Groningen,
The Netherlands
Erik R. Ranschaert, Department of Radiology, ETZ Hospital,
Tilburg, The Netherlands
Annalisa Trianni, Department of Medical Physics, ASUIUD,
UDINE, Udine, Italy
Michail E. Klontzas, University Hospital of Heraklion, Heraklion,
Greece
Institute of Computer Science, Foundation for Research and
Technology (FORTH), Heraklion, Greece
The series Imaging Informatics for Healthcare Professionals is the
ideal starting point for physicians and residents and students in
radiology and nuclear medicine who wish to learn the basics in
different areas of medical imaging informatics. Each volume is
a short pocket-sized book that is designed for easy learning and
reference.
The scope of the series is based on the Medical Imaging
Informatics subsections of the European Society of Radiology
(ESR) European Training Curriculum, as proposed by ESR and
the European Society of Medical Imaging Informatics (EuSoMII).
The series, which is endorsed by EuSoMII, will cover the curric-
ula for Undergraduate Radiological Education and for the level
I and II training programmes. The curriculum for the level III
training programme will be covered at a later date. It will offer
frequent updates as and when new topics arise.
Michail E. Klontzas •
Salvatore Claudio Fanni •
Emanuele Neri
Editors
Introduction to
Artificial Intelligence
Editors
Michail E. Klontzas Salvatore Claudio Fanni
University Hospital of Heraklion Academic Radiology, Department
Heraklion, Greece of Translational Research
University of Pisa
Institute of Computer Science,
Pisa, Pisa, Italy
Foundation for Research and
Technology (FORTH)
Heraklion, Greece
Emanuele Neri
Academic Radiology, Department
of Translational Research
University of Pisa
Pisa, Italy
© EuSoMII 2023
This work is subject to copyright. All rights are solely and exclusively licensed by
the Publisher, whether the whole or part of the material is concerned, specifically
the rights of translation, reprinting, reuse of illustrations, recitation, broadcasting,
reproduction on microfilms or in any other physical way, and transmission or
information storage and retrieval, electronic adaptation, computer software, or by
similar or dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks,
etc. in this publication does not imply, even in the absence of a specific statement,
that such names are exempt from the relevant protective laws and regulations and
therefore free for general use.
The publisher, the authors, and the editors are safe to assume that the advice
and information in this book are believed to be true and accurate at the date of
publication. Neither the publisher nor the authors or the editors give a warranty,
expressed or implied, with respect to the material contained herein or for any errors
or omissions that may have been made. The publisher remains neutral with regard
to jurisdictional claims in published maps and institutional affiliations.
v
vi Preface
vii
viii Contents
E. Koltsakis ()
Department of Radiology, Karolinska University Hospital, Stockholm,
Sweden
M. E. Klontzas
University Hospital of Heraklion, Heraklion, Greece
Institute of Computer Science, Foundation for Research and Technology
(FORTH), Heraklion, Greece
A. H. Karantanas
Department of Medical Imaging, University Hospital of Heraklion,
Heraklion, Crete, Greece
Department of Radiology, School of Medicine, University of Crete,
Heraklion, Crete, Greece
Advanced Hybrid Imaging Systems, Institute of Computer Science,
Foundation for Research and Technology (FORTH), Heraklion, Crete,
Greece
(continued)
6 E. Koltsakis et al.
The edges and the neurons are weighted and the weights
can be adjusted through the learning process of the
machine. As long as the output is not the desired and
there is a difference (error) the machine will adjust the
weight of the neuron in order to reduce that error. The
neurons are arranged in layers. The first layer corre-
sponds the input layer and the last to the output.
As the research usage of AI, ML, DL, and CNN became greater
and greater, new milestones were added to the timeline. In 2002,
Torch, the first machine learning library that provided algo-
rithms for DNNs, was created and released by Ronan Collobert,
Samy Bengio, and Johnny Mariéthoz. As the name implies,
in a machine learning library, one may find common learning
algorithms that are available for the public. Depending on the
purpose of the program or the programming language, different
libraries are more applicable than others, similar to traditional
libraries. In the following years, important changes escalated
quickly.
In 2005, a Stanford robot won the DARPA Grand Challenge by
driving autonomously for 131 miles along an unrehearsed desert
trail. With raw data from GPS, camera, and 3D mapping com-
posed with LIDARS, the car controlled the speed and direction in
order to avoid obstacles.
In 2006, the term “Machine Reading” was used by Oren
Etzioni, Michele Banko, Michael J. Cafarella to describe the
8 E. Koltsakis et al.
References
1. Turing AM. Computing machinery and intelligence. Mind.
1950;LIX:433–60.
2. Strachey CS. Logical or non-mathematical programmes, ACM ’52, 1952.
3. Weizenbaum J. ELIZA-A computer program for the study of natural
language communication between man and machine. Commun ACM.
1966;9:36–45.
4. Samuel AL. Some studies in machine learning. IBM J Res Dev.
1959;3:210–29.
5. Michie D. Experiments on the mechanization of game-learning. Part I.
Characterization of the model and its parameters. Comput J. 1963;6:232–
6.
6. Shortlife EH, Buchanan BG. A model of inexact reasoning in medicine.
Math Biosci. 1975;23:351–79.
7. Becker S, Hinton GE. Self-organizing neural network that discovers
surfaces in random-dot stereograms. Nature. 1992;355:161–3.
8. Fukushima K. Neocognitron: a self-organizing neural network model for
a mechanism of pattern recognition unaffected by shift in position. Biol
Cybern. 1980;36:193–202.
9. Hubel DH, Wiesel TN. Receptive fields of single neurones in the cat’s
striate cortex. J Physiol. 1959;148:574–91.
10. Lecun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied
to document recognition. Proc IEEE. 1998;86:2278–324.
11. Hinton G. Deep learning-a technology with the potential to transform
health care. JAMA. 2018;320:1101–2.
12. Raymond GJ. Ethics of artificial intelligence in radiology: summary
of the joint European and North American Multisociety Statement.
Radiology. 2019;293:436–40.
Using Commercial and Open-Source
Tools for Artificial Intelligence: A 2
Case Demonstration on a Complete
Radiomics Pipeline
2.1 Introduction
A. Matikas · T. Foukakis
Department of Oncology-Pathology, Karolinska Institutet, Stockholm,
Sweden
G. C. Manikis ()
Computational BioMedicine Laboratory, Institute of Computer Science,
Foundation for Research and Technology (FORTH), Heraklion, Greece
Department of Oncology-Pathology, Karolinska Institutet, Stockholm,
Sweden
2 Commercial and Open-Source Tools for AI 15
It is evident from Table 2.1 that no single software tool can serve
as a one-stop-shop solution for the design of a complete radiomics
analysis workflow since each one has its own design style,
customization, and functionality. To the best of our knowledge,
several AI models can be deployed and executed within the
3DSlicer ecosystem (e.g., MONAI), and however this integration
demands Python software skills to enable compatibility between
the components. The user lacking programming skills can only
partially exploit the capabilities of each software. It might be
discouraging to create a seamless AI workflow where the output
of each action is fed as input to a different software platform, not
only because a general overview of the available tools is necessary
but also because ensuring compatibility is a non-trivial task. A
possible course of actions to structure a complete ML-based
radiomics project in practice is presented in the following para-
graph, shedding light on the interplay between different actions
and how they can be combined to compose the full pathway from
the clinical question to the AI derived answer. To this direction,
our demonstration on how to perform a radiomics workflow starts
with an MRI region of a cancer patient to predict diagnosis
and assess the evolution of cancer and its mortality. Since our
proposed pipeline is designed to be used by doctors or medical
physicists with little or no programming skills, we recommend
the 3DSlicer as a user-friendly tool which provides the flexibility
on using the PyRadiomics benefits for both image pre-processing
and radiomics feature extraction steps. As for the radiomics
harmonization process, ComBaTool seems to be the ideal tool as it
needs no programming skills and can be used online by selecting
each parameter manually. Then, in order to develop the ML-based
2 Commercial and Open-Source Tools for AI 23
B. Image pre-processing
with 3DSlicer & PyRadiomics
Intensity Normalization
Reconstruction / Registration
Discretization
E. Radiomics Modeling
D. Radiomics Harmonization
with RapidMiner C. Radiomics extraction
with ComBaTool
with 3DSlicer & PyRadiomics
radiomics features
deep features
Fig. 2.1 A proposed radiomics analysis pipeline using commercial tools and plugins
E. Stamoulou et al.
2 Commercial and Open-Source Tools for AI 27
2.7 Discussion
2.8 Conclusion
References
1. Lambin P, Leijenaar RTH, Deist TM, Peerlings J, De Jong EEC,
Van Timmeren J, Sanduleanu S, Larue RTHM, Even AJG, Jochems A,
et al. Radiomics: the bridge between medical imaging and personalized
medicine. Nat Rev Clin Oncol. 2017;14(12):749–62
30 E. Stamoulou et al.
54. Bazin P-L, Cuzzocreo JL, Yassa MA, Gandler W, McAuliffe MJ,
Bassett SS, Pham DL. Volumetric neuroimage analysis extensions for
the MIPAV software package. J Neurosci Methods. 2007;165(1):111–
21
55. Götz M, Nolden M, Maier-Hein K. MITK phenotyping: an open-
source toolchain for image-based personalized medicine with radiomics.
Radiother Oncol. 2019;131:108–11
56. Nioche C, Orlhac F, Boughdad S, Reuzé S, Goya-Outi J, Robert C,
Pellot-Barakat C, Soussan M, Frouin F, Buvat I. LIFEx: a freeware
for radiomic feature calculation in multimodality imaging to accelerate
advances in the characterization of tumor heterogeneity. Cancer Res.
2018;78(16):4786–9
57. Yaniv Z, Lowekamp BC, Johnson HJ, Beare R. Simpleitk image-
analysis notebooks: a collaborative environment for education and
reproducible research. J Digital Imaging. 2018;31(3):290–303
58. Van Griethuysen JJM, Fedorov A, Parmar C, Hosny A, Aucoin N,
Narayan V, Beets-Tan RGH, Fillion-Robin J-C, Pieper S, Aerts HJWL.
Computational radiomics system to decode the radiographic phenotype.
Cancer Res. 2017;77(21):e104–7
59. van Timmeren JE, Cester D, Tanadini-Lang S, Alkadhi H, Baessler B.
Radiomics in medical imaging—“how-to” guide and critical reflection.
Insights Imaging. 2020;11:91
60. Szczypiński PM, Strzelecki M, Materka A, Klepaczko A. Mazda–
the software package for textural analysis of biomedical images. In:
Computers in medical activity. Berlin: Springer; 2009. p. 73–84.
61. Zhang L, Fried DV, Fave XJ, Hunter LA, Yang J, Court LE. Ibex: an
open infrastructure software platform to facilitate collaborative work in
radiomics. Med Phys. 2015;42(3):1341–53
62. RadiomiX Research Toolbox. https://radiomics.bio/radiomix-toolbox/.
Available online; accessed on 21 Nov 2022
63. Talukder S. GPU-based medical image segmentation: Brain MRI analy-
sis using 3d slicer. In: Artificial intelligence applications for health care.
Boca Raton: CRC Press; 2022. p. 109–121
64. Deep Learning Studio. https://deeplearningstudio.com/. Available
online; accessed on 21 Nov 2022
65. Nvidia’s Digits System. https://developer.nvidia.com/digits. Available
online; accessed on 21 Nov 2022
66. Zwanenburg A, Vallières M, Abdalah MA, Aerts HJWL, Andrearczyk
V, Apte A, Ashrafinia S, Bakas S, Beukinga RJ, Boellaard R, et al. The
image biomarker standardization initiative: standardized quantitative
radiomics for high-throughput image-based phenotyping. Radiology.
2020;295(2):328–38
67. Zwanenburg A, Leger S, Vallières M, Löck S. Image biomarker stan-
dardisation initiative; arxiv 2016. Preprint. arXiv:1612.07003
2 Commercial and Open-Source Tools for AI 35
68. Fortin J-P, Cullen N, Sheline YI, Taylor WD, Aselcioglu I, Cook PA,
Adams P, Cooper C, Fava M, McGrath PJ, et al. Harmonization of
cortical thickness measurements across scanners and sites. Neuroimage.
2018;167:104–20
69. Garcia-Dias R, Scarpazza C, Baecker L, et al. Neuroharmony: A new
tool for harmonizing volumetric MRI data from unseen scanners. Neu-
roimage. 2020;220:117127. https://doi.org/10.1016/j.neuroimage.2020.
117127
70. Bouhali O, Bensmail H, Sheharyar A, David F, Johnson JP. A review of
radiomics and artificial intelligence and their application in veterinary
diagnostic imaging. Vet Sci. 2022;9(11):620
71. Wagner MW, Namdar K, Biswas A, Monah S, Khalvati F, Ertl-Wagner
BB. Radiomics, machine learning, and artificial intelligence—what the
neuroradiologist needs to know. Neuroradiology. 2021;63(12):1957–67
72. Webb GI, Keogh E, Miikkulainen R. Naïve bayes. Encycl Mach Learn.
2010;15:713–14
73. Wright RE, Logistic regression. In: Grimm LG, Yarnold PR, editors.
Reading and understanding multivariate statistics. American Psycholog-
ical Association; 1995. p. 217–44
74. Suthaharan S. Support vector machine. In: Machine learning models and
algorithms for big data classification. Berlin: Springer; 2016. p. 207–35
75. Myles AJ, Feudale RN, Liu Y, Woody NA, Brown SD. An introduction
to decision tree modeling. J Chemom J Chemom Soc. 2004;18(6):275–
85
76. Breiman L. Random forests. Mach Learn. 2001;45(1):5–32
77. Puttagunta M, Ravi S. Medical image analysis based on deep learning
approach. Multimedia Tools Appl. 2021;80(16):24365–98
78. Hodgson J. The 5 Stages of Machine Learning Validation. Retrieved
from https://towardsdatascience.com/the-5-stages-of-machine-
learning-validation-162193f8e5db, 2022
79. Karthikeyan N. Step-by-step guide for a deep learning project.
Retrieved from https://medium.com/@neelesh_k/structuring-deep-
learning-projects-b83d29513aea, 2022
80. Brownlee, J. Machine learning mastery with python: understand your
data, create accurate models, and work projects end-to-end; 2016.
Melbourne: Machine Learning Mastery
81. Hutter F, Kotthoff L, Vanschoren J. Automated machine learning:
methods, systems, challenges; 2019. Berlin: Springer Nature
82. Mustafa A, Azghadi MR. Automated machine learning for healthcare
and clinical notes analysis. Computers. 2021;10(2):24
83. Microsoft AutoMl. https://www.microsoft.com/en-us/research/project/
automl/. Available online; accessed on 21 Nov 2022
84. Goudas T, Doukas C, Chatziioannou A, Maglogiannis I. A collabo-
rative biomedical image-mining framework: application on the image
analysis of microscopic kidney biopsies. IEEE J Biomed Health Inf.
2012;17(1):82–91
36 E. Stamoulou et al.
3.1 Introduction
Fig. 3.3 Three steps of Supervised Learning: (i) training of the algorithm,
(ii) validation of the trained model, and (iii) test of the model
Fig. 3.4 How classification (on the left) and regression algorithms (on the
right) work. Classification algorithms find out the better hyperplane(s) which
divides the data into two (or more) classes; a regression model aims to find
out the better function that approximates the trend of the data
Fig. 3.5 How Unsupervised Machine Learning algorithms work. They use a
more self-contained approach, in which a computer learns to identify patterns
without any guidance, only inputting data that are unlabeled and for which no
specific output has been defined. In practice, this type of algorithms will learn
to divide the data into different clusters based on the characteristics that unite
or discriminate them the most
3.5 Conclusions
USEFUL GLOSSARY
References
1. Wartman S, Combs C. Medical education must move from the informa-
tion age to the age of artificial intelligence. Acad Med. 2018;93:1107–9.
2. Obermeyer Z, Lee T. Lost in thought – the limits of the human mind and
the future of medicine. N Engl J Med. 2017;377(13):1209–11.
3. Vamathevan J, Clark D, Czodrowski P, Dunham I, Ferran E, Lee G, Li
B, Madabhushi A, Shah P, Spitzer M, Zhao S. Applications of machine
learning in drug discovery and development. Nat Rev Drug Discov.
2019;18:463–77.
4. Valletta JJ, Torney C, Kings M, Thornton A, Madden J. Applica-
tions of machine learning in animal behaviour studies. Anim Behav.
2017;124:203–20.
5. Recknagel F. Applications of machine learning to ecological modelling.
Ecol Modell. 2001;146:303–310.
6. Garg A, Mago V. Role of machine learning in medical research: a survey.
Comput Sci Rev. 2021;40:100370.
7. Rajkomar A, Dean J, Kohane I. Machine learning in medicine. N Engl J
Med. 2019;380:1347–1358.
3 Introduction to Machine Learning in Medicine 61
4.1 Introduction
M. E. Klontzas ()
University Hospital of Heraklion, Heraklion, Greece
Institute of Computer Science, Foundation for Research and Technology
(FORTH), Heraklion, Greece
R. Cuocolo
Department of Medicine, Surgery and Dentistry, University of Salerno,
Baronissi, Italy
Fig. 4.1 Overview of the most common machine learning methods used in
radiomics analysis (created with biorender.com)
4.2.1 R-CNN
4.2.3 DeepLab
4.3.2 Clustering
4.4.1 Boruta
Once the final set of important features has been selected using
the aforementioned methods the final step is to use these features
to create predictive models. These can be classification or regres-
sion models and a set of traditional machine learning methods
can be used for this purpose. The most commonly used ones
are logistic regression, decision trees, random forests, gradient
boosting models, support vector machines (SVM), and neural
80 M. E. Klontzas and R. Cuocolo
Neural networks have also been used with radiomics data but will
be extensively discussed in Chap. 6. However, it is important to
mention that even though deep learning is excellent in computer
vision tasks, it has been proven to underperform when used with
tabular data, especially given the relatively small size of datasets
typically available in medical imaging. Importantly it has been
shown that when using tabular data, methods such as ensembles
(e.g., random forest and XGBoost) may perform better than deep
learning models [39] and need less data and tuning than deep
learning models. This is the reason why other machine learning
algorithms are often preferred over deep learning when creating
radiomics predictive models in literature.
4.6 Conclusion
References
1. van Timmeren JE, Cester D, Tanadini-Lang S, Alkadhi H, Baessler B.
Radiomics in medical imaging—“how-to” guide and critical reflection.
Insights Imaging. 2020;11:91.
2. Yamada R, Okada D, Wang J, Basak T, Koyama S. Interpretation of omics
data analyses. J Human Gen. 2021;66(1):93–102. https://doi.org/10.1038/
s10038-020-0763-5.
3. Gillies RJ, Kinahan PE, Hricak H. Radiomics: images are more than
pictures, they are data. Radiology. 2016;278(2):563–77. [cited 2021 Oct
18] https://pubs.rsna.org/doi/abs/10.1148/radiol.2015151169
4. Rohart F, Gautier B, Singh A, Lê Cao KA. mixOmics: an R package
for ‘omics feature selection and multiple data integration. PLoS Comput
Biol. 2017;13(11):e1005752.
5. Lambin P, Leijenaar RTH, Deist TM, Peerlings J, de Jong EEC, van
Timmeren J, et al. Radiomics: the bridge between medical imaging and
personalized medicine. Nat Rev Clin Oncol. 2017;14(12):749–62. https:/
/doi.org/10.1038/nrclinonc.2017.141.
6. Papanikolaou N, Matos C, Koh DM. How to develop a meaningful
radiomic signature for clinical use in oncologic patients. Cancer Imaging.
2020;20(1):33.
7. Girshick R, Donahue J, Darrell T, Malik J. Rich feature hierar-
chies for accurate object detection and semantic segmentation. ArXiv.
2013;1311.2524. http://arxiv.org/abs/1311.2524.
8. Girshick R. Fast R-CNN. ArXiv. 2015;1504.08083. http://arxiv.org/abs/
1504.08083.
9. He K, Gkioxari G, Dollár P, Girshick R. Mask R-CNN. ArXiv.
2017;1703.06870. http://arxiv.org/abs/1703.06870.
10. Ronneberger O, Fischer P, Brox T. U-Net: convolutional networks
for biomedical image segmentation. ArXiv. 2015;1505.04597. http://
arxiv.org/abs/1505.04597.
84 M. E. Klontzas and R. Cuocolo
27. Li ZD, Guo W, Ding SJ, Chen L, Feng KY, Huang T, et al. Identifying
key microRNA signatures for neurodegenerative diseases with machine
learning methods. Front Genet. 2022;13:880997.
28. Zhang YH, Li H, Zeng T, Chen L, Li Z, Huang T, et al. Identifying
transcriptomic signatures and rules for SARS-CoV-2 infection. Front Cell
Dev Biol. 2021;8:627302.
29. Wu X, Kumar V, Ross QJ, Ghosh J, Yang Q, Motoda H, et al. Top 10
algorithms in data mining. Knowl Inf Syst. 2008;14(1):1–37.
30. Goldstein BA, Polley EC, Briggs FBS. Random forests for genetic
association studies. Stat Appl Genet Mol Biol. 2011;10(1):32.
31. Toloşi L, Lengauer T. Classification with correlated features: unreliability
of feature ranking and solutions. Bioinformatics. 2011;27(14):1986–94.
32. Freund Y, Schapire RE. A decision-theoretic generalization of on-line
learning and application to boosting. J Comput Syst Sci. 1997;55:119–
39.
33. He Z, Lin D, Lau T, Wu M. Gradient boosting machine: a survey point
zero one technology. ArXiv. 2019;1908.06951:1–9.
34. Ruder S. An overview of gradient descent optimization algorithms.
ArXiv. 2016;1609.04747:1–14. http://arxiv.org/abs/1609.04747.
35. Chen T, Guestrin C. XGBoost: a scalable tree boosting system. ArXiv.
2016;1603.02754:1–13. http://arxiv.org/abs/1603.02754.
36. Klontzas ME, Manikis GC, Nikiforaki K, Vassalou EE, Spanakis K,
Stathis I, et al. Radiomics and machine learning can differentiate
transient osteoporosis from avascular necrosis of the hip. Diagnostics.
2021;11:1686.
37. Chen PT, Chang D, Yen H, Liu KL, Huang SY, Roth H, et al. Radiomic
features at CT can distinguish pancreatic cancer from noncancerous
pancreas. Radiol Imaging Cancer. 2021;3(4):e210010.
38. Awe AM, van den Heuvel MM, Yuan T, Rendell VR, Shen M, Kampani
A, et al. Machine learning principles applied to CT radiomics to predict
mucinous pancreatic cysts. Abdom Radiol. 2022;47(1):221–31.
39. Shwartz-Ziv R, Armon A. Tabular data: deep learning is not all you need.
ArXiv. 2021;2106.03253:1–13. http://arxiv.org/abs/2106.03253.
Natural Language Processing
5
Salvatore Claudio Fanni, Maria Febi,
Gayane Aghakhanyan, and Emanuele Neri
there were still no computers that came anywhere near being able
to carry on a basic conversation. The machine translation research
was almost killed by the 1966 ALPAC Report, which concluded
that machine translation was nowhere near achievement and led
to significant cut of funding [4].
The second phase of NLP was prompted by AI, with much
more emphasis on world knowledge and on its role in the con-
struction and manipulation of meaning representations. Overall, it
took nearly 14 years (until 1980) for NLP AI research to recover
from the broken expectations created by extreme enthusiasts
during the first phase of the development. Albeit the second
phase of NLP work was AI-driven and semantics-oriented, the
third phase can be described, in reference to its dominant style,
as a grammatical-logical phase. This trend, as a response to
the failures of practical system building, was stimulated by the
development of grammatical theory among linguists during the
1970s, and by the move toward the use of logic for knowledge
representation and reasoning in AI. Computational grammar the-
ory became a very active area of research linked with work on
logics for meaning and knowledge representation that can deal
with the language user’s beliefs and intentions and can capture
discourse features and functions like emphasis and theme, as
well as indicate semantic case roles. Research and development
extended worldwide, notably in Europe and Japan, aimed not only
at interface subsystems but at autonomous NLP systems, as for
message processing or translation [1].
Until the 1980s, the majority of NLP systems used complex,
“handwritten” rules, however, in the late 1980s, a revolution in
NLP came about. This was the result of both the steady increase
of computational power, and the shift to machine learning (ML)
algorithms. In the 1990s, the popularity of statistical models
for natural language processes analyses rose dramatically. The
pure statistics NLP methods have become remarkably valuable.
In addition, the recurrent neural network (RNN) models have
been introduced and found their niche in 2007 for voice and
text processing. Currently, neural network models are considered
the cutting edge of research and development in the NLP’s
understanding of text and speech generation [1].
90 S. C. Fanni et al.
References
1. Kochmar E. Getting started with natural language processing. New York:
Simon and Schuster; 2022.
2. Turing AM. Computing machinery and intelligence. Mind.
1950;236:433–60.
3. Harnad S. Minds, machines and searle. J Exp Theor Artif Intell.
1989;1(1):5–25.
5 Natural Language Processing 97
18. Lindberg C. The unified medical language system (UMLS) of the national
library of medicine. J Am Med Rec Assoc. 1990;61(5):40–2.
19. Millar J. The need for a global language – SNOMED CT introduction.
Stud Health Technol Inform. 2016;225:683–5.
20. Khurana D, Koli A, Khatter K, Singh S. Natural language processing:
state of the art, current trends and challenges. Multimed Tools Appl. 2022.
https://doi.org/10.1007/s11042-022-13428-4.
21. Garg P, Girdhar N. A systematic review on spam filtering tech-
niques based on natural language processing framework. In: 2021
11th international conference on cloud computing, data science
& engineering (confluence). 2021. p. 30–5. https://doi.org/10.1109/
Confluence51648.2021.9377042.
22. Malmasi S, Hosomura N, Chang L-S, Brown CJ, Skentzos S, Turchin A.
Extracting healthcare quality information from unstructured data. AMIA
Annu Symp Proc. 2017;2017:1243–52.
23. Iroju OG, Olaleke JO. A systematic review of natural language processing
in healthcare. BMC Med Inform Decis Mak. 2015;21:179. https://doi.org/
10.5815/ijitcs.2015.08.07.
24. Jain NL, Knirsch CA, Friedman C, Hripcsak G. Identification of sus-
pected tuberculosis patients based on natural language processing of
chest radiograph reports. In: Proceedings: a Conference of the American
Medical Informatics Association. AMIA Fall Symposium. 1996. pp. 542–
6.
25. Jain NL, Friedman C. Identification of findings suspicious for breast
cancer based on natural language processing of mammogram reports.
Proceedings: a Conference of the American Medical Informatics Asso-
ciation. AMIA Fall Symposium. 1997. pp. 829–33.
26. Savova GK, et al. Mayo clinical text analysis and knowledge extraction
system (cTAKES): architecture, component evaluation and applications.
J Am Med Inform Assoc. 2010;17(5):507–13. https://doi.org/10.1136/
jamia.2009.001560.
27. Panny A, et al. A methodological approach to validate pneumonia
encounters from radiology reports using natural language processing.
Methods Inf Med. 2022;61(1–2):38–45. https://doi.org/10.1055/a-1817-
7008.
28. Jackson R, et al. CogStack – experiences of deploying integrated infor-
mation retrieval and extraction services in a large National Health Service
Foundation Trust hospital. BMC Med Inform Decis Mak. 2018;18(1):1–
13. https://doi.org/10.1186/s12911-018-0623-9.
29. Wang T, et al. Implementation of a real-time psychosis risk detection and
alerting system based on electronic health records using CogStack. J Vis
Exp. 2020;159:60794. https://doi.org/10.3791/60794.
5 Natural Language Processing 99
Abbreviations
AE autoencoder
AI Artificial intelligence
ANN Artificial neural networks
CAD/CADe Computer-aided diagnosis/detection
CLAHE Contrast-limited adaptive histogram equalization
CNN Convolutional neural networks
CT Computed tomography
DL Deep learning
DrCNN Denoising residual convolutional neural network
FL Federated learning
E. Trivizakis ()
Computational BioMedicine Laboratory (CBML), Institute of Computer
Science (ICS), Foundation for Research and Technology-Hellas (FORTH),
Heraklion, Greece
e-mail: [email protected]
K. Marias
Computational BioMedicine Laboratory (CBML), Institute of Computer
Science (ICS), Foundation for Research and Technology-Hellas (FORTH),
Heraklion, Greece
Department of Electrical and Computer Engineering, Hellenic
Mediterranean University, Heraklion, Greece
e-mail: [email protected]
The field of medical image analysis has recently shifted its focus
from traditional “hand-crafted” image processing and simple
statistical models to the cutting-edge technique of deep learning
analysis. Medical professionals can be potentially benefited by
the accurate lesion identification, segmentation of regions of
interest, progression tracking, and categorization of pathological
anatomical structures to aid them in clinical practice. Therefore,
it is crucial for healthcare to adopt DL-based applications for
the aforementioned tasks, since they can provide overburdened
doctors with more agency and facilitate swift decision-making in
the highly demanding clinical environment.
The data analysis in deep neural networks follows a hierar-
chical architecture that progressively identifies hidden patterns
inside the examined region of interest [1] and can potentially
correlate those with clinical outcomes. Like biological neurons,
artificial neurons receive a number of inputs, perform some sort
of computation, and then output the results. In each neuron, a
straightforward computation is performed, including a nonlinear
activation function and a mechanism for sparse feature coding.
Some typical nonlinear activation [2] functions of ANN include
the sigmoid transformation, hyperbolic tangent, and the com-
6 Deep Learning Fundamentals 103
Fig. 6.1 Deep learning architectures: (a) convolutional neural network, (b)
artificial neural network, (c) autoencoder
6.3.5 Harmonization
Fig. 6.4 The architecture of a DrCNN model used for denoising. This model
architecture estimates the noise distribution patterns of the input dataset.
DrCNN, denoising residual convolutional neural network
Fig. 6.5 Two patch extraction methods are presented: (a) exhaustive with no
overlapping patches from a high resolution pathology image and (b) based on
regions of interest
Fig. 6.6 The two types of transfer learning methods that have been proposed
in the literature: (a) fine-tuning TL, the transferred weights are adapted for
the new data distribution, and (b) “off-the-shelf” TL, only the convolutional
weights are transferred to the target model for feature extraction
6.5.1 Reproducibility
6.5.2 Traceability
6.5.3 Explainability
6.5.4 Trustworthiness
References
1. Chen X, Wang X, Zhang K, Fung KM, Thai TC, Moore K, Mannel RS,
Liu H, Zheng B, Qiu Y. Recent advances and clinical applications of deep
learning in medical image analysis. Med Image Anal. 2022;79:102444.
2. He K, Zhang X, Ren S, Sun J. Delving deep into rectifiers: surpassing
human-level performance on ImageNet classification. 2015.
3. Rumelhart DE, Hinton GE, Williams RJ. Learning representations by
back-propagating errors. Nat 1986 3236088. 1986;323:533–6.
4. Jaiswal A, Babu AR, Zadeh MZ, Banerjee D, Makedon F. A survey on
contrastive self-supervised learning. Technol 2021. 2020;9:2.
5. Luca AR, Ursuleanu TF, Gheorghe L, Grigorovici R, Iancu S, Hlusneac
M, Grigorovici A. Impact of quality, type and volume of data used by
deep learning models in the analysis of medical images. Informatics Med
Unlocked. 2022;29:100911.
6. Xia W, Hu B, Li H, et al. Deep Learning for automatic differential
diagnosis of primary central nervous system lymphoma and glioblastoma:
multi-parametric magnetic resonance imaging based convolutional neural
network model. J Magn Reson Imaging. 2021;54:880–7.
7. Trivizakis E, Manikis GC, Nikiforaki K, Drevelegas K, Constantinides M,
Drevelegas A, Marias K. Extending 2D convolutional neural networks to
3D for advancing deep learning cancer classification with application to
MRI liver tumor differentiation. IEEE J Biomed Heal Inform. 2018:1–1.
8. Asuntha A, Srinivasan A. Deep learning for lung Cancer detection and
classification. Multimed Tools Appl. 2020;79:7731–62.
9. Trivizakis E, Ioannidis G, Melissianos V, Papadakis G, Tsatsakis A,
Spandidos D, Marias K. A novel deep learning architecture outper-
forming ‘off-the-shelf’ transfer learning and feature-based methods in
128 E. Trivizakis and K. Marias
7.1 Introduction
Fig. 7.1 Overview of the impact of data quality and numerosity in clinical
image analysis using AI
Fig. 7.2 The medical imaging formation process. The final image is the
result of different processes related to image acquisition and processing
uniformity, etc. While these indices are extremely useful and valid
for describing the intrinsic characteristics of the image, they need
to be supported by other information that takes into account the
“diagnostic” quality of a medical image.
The diagnostic quality of a medical image is defined as “the
quality of decisions taken as a result of the image evaluation in
terms of benefit for the patient (does the therapeutic treatment
change? What is the effect on mortality? etc.)” [24]. Clearly, a
radiologist does not necessarily need a “beautiful” image (assum-
ing he can define what the term “beautiful” means), but rather the
image that provides the most useful and effective representation
for his purposes, such as the identification and classification of a
lesion, or the pathological condition, etc. Moreover, with the same
diagnostic quality, thanks to his experience, the clinical context,
and the capabilities of the human mind, a radiologist is able to
respond to the specific task even with two images of different
intrinsic quality. Of course, it remains clear that diagnostic quality
is more difficult to achieve in highly complex clinical questions,
looking for small structural or functional changes, for example.
Therefore, in radiology, since human observers are the final
recipients of visual information, subjective evaluation of image
quality is currently considered the most reliable approach [26].
Fig. 7.3 An example of the impact of data quality and numerosity on the
performance of a machine learning algorithm for a classification task. Each
point represents a patient with a colour referring to three different pathologies
(yellow, green, and purple). The sketch shows that, in the presence of few
data with low quality, the algorithm is unable to separate the pathologies. By
simultaneously improving the quality and number of data, the algorithm’s
performance increases
since, on the one hand, each image of the dataset has sufficient
diagnostic quality to solve the problem, and, on the other hand,
there are be no sources of unwanted variability between the
different images, due to differences in intrinsic quality, that could
confuse the AI algorithms.
In medical imaging, acquiring a set of images with intrinsic
“constant” quality means examining all the subjects in the same
setting, with the same scanner, the same acquisition protocol,
carrying out the same preprocessing, etc. In this scenario, it is
not easy to obtain a large dataset. Therefore, in recent years,
many studies have combined data and images collected in dif-
ferent ways (different acquisition institutes, scanners, acquisition
protocols, etc.) to obtain a multicentre dataset that is clinically
representative of the population to be analysed. Thereby, each
image showed sufficient diagnostic quality but different intrinsic
quality as a function of the different acquisition protocols and
processing parameters. Therefore, the process of appropriately
combining data from different sources, known as data pooling,
is becoming fundamental to the success of AI in radiology and
140 A. Barucci et al.
Standardisation
Standardisation of datasets is a common requirement for many
machine learning models that might behave badly if individual
features show different ranges. One way to standardise the data
is to remove the mean value from each feature and divide it by
the standard deviation. The mean value and standard deviation are
calculated across samples. Other examples of data standardisation
are detailed in reference [33].
Dimensionality Reduction
Dimensionality reduction refers to the process of reducing the
number of features in a dataset while keeping as much variation
in the original dataset as possible. Dimensionality reduction could
manage multicollinearity of the features and remove noise in the
data. From a machine learning analysis perspective, a lower num-
ber of features means less training time and less computational
power. It also avoids the potential problem of overfitting, lead-
ing to an increase in overall performance. Principal component
analysis (PCA) is a linear dimensionality reduction technique that
transforms a set of correlated variables into a smaller number
of uncorrelated variables, called principal components, while
retaining as much variation in the original dataset as possible [53].
Other linear dimensionality reduction methods are factor analysis
(FA) [54] and linear discriminant analysis (LDA) [55].
Feature Selection
In machine learning and statistics, feature selection is the pro-
cess of selecting a subset of relevant features for use in model
construction. In medicine and health care, feature selection is
advantageous because it enables the interpretation of the machine
learning model and the discovery of new potential biomarkers
related to a specific disorder or condition [56]. Feature selection
methods can be grouped into three categories: filter method,
wrapper method, and embedded method [57, 58]. In the filter
method, features are selected based on the general characteristics
of the dataset without using any predictive model. In the wrapper
7 Data Preparation for AI Analysis 143
References
1. Barucci A, Cucci C, Franci M, Loschiavo M, Argenti F. A deep learning
approach to ancient Egyptian hieroglyphs classification. IEEE Access.
2021;9:123438–47.
2. Cucci C, Barucci A, Stefani L, Picollo M, Jiménez-Garnica R, Fuster-
Lopez L. Reflectance hyperspectral data processing on a set of Picasso
paintings: which algorithm provides what? A comparative analysis of
multivariate, statistical and artificial intelligence methods. In: Groves R,
Liang H, editors. Optics for arts, architecture, and archaeology VIII.
Bellingham: SPIE; 2021. p. 1.
3. Li Z, Shen H, Cheng Q, Liu Y, You S, He Z. Deep learning based
cloud detection for medium and high resolution remote sensing images
of different sensors. ISPRS J Photogramm Remote Sens. 2019;150:197–
212.
4. Scapicchio C, Gabelloni M, Barucci A, Cioni D, Saba L, Neri E. A deep
look into radiomics. Radiol Med. 2021;126(10):1296–311.
5. Hamet P, Tremblay J. Artificial intelligence in medicine. Metabolism.
2017;69:36–40.
6. Jordan MI, Mitchell TM. Machine learning: trends, perspectives, and
prospects. Science. 2015;349(6245):255–60.
7. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature.
2015;521(7553):436–44.
8. Ismael AM, Şengür A. Deep learning approaches for COVID-19 detec-
tion based on chest X-ray images. Expert Syst Appl. 2021;164:114054.
9. Ding Y, Sohn JH, Kawczynski MG, Trivedi H, Harnish R, Jenkins NW,
et al. A deep learning model to predict a diagnosis of Alzheimer disease
by using 18F-FDG PET of the brain. Radiology. 2019;290(2):456–64.
10. Van Sloun RJ, Cohen R, Eldar YC. Deep learning in ultrasound imaging.
Proc IEEE. 2019;108(1):11–29.
11. Deng S, Zhang X, Yan W, Chang EI, Fan Y, Lai M, et al. Deep learning in
digital pathology image analysis: a survey. Front Med. 2020;14(4):470–
87.
12. Guiot J, Vaidyanathan A, Deprez L, Zerka F, Danthine D, Frix A, et al. A
review in radiomics: making personalized medicine a reality via routine
imaging. Med Res Rev. 2022;42(1):426–40.
13. Consortium TM. Project MONAI. Zenodo; 2020. https://zenodo.org/
record/4323059.
14. van Leeuwen KG, Schalekamp S, Rutten MJCM, van Ginneken B, de
Rooij M. Artificial intelligence in radiology: 100 commercially available
products and their scientific evidence. Eur Radiol. 2021;31(6):3797–804.
15. imbio. https://www.imbio.com.
16. BRAINOMIX. https://www.brainomix.com.
17. Goebel J, Stenzel E, Guberina N, Wanke I, Koehrmann M, Kleinschnitz
C, et al. Automated ASPECT rating: comparison between the Frontier
7 Data Preparation for AI Analysis 147
8.1 Introduction
8.2 Detection
8.3 Classification
8.4 Segmentation
include brain [36], liver [37], lung [38], breast [39], head and neck
[40], rectum [41], and stomach [42].
8.4.1 Monitoring
8.4.2 Prediction
accuracy and patient care; on the other hand, the radiology field
is not exempt from the inverse relationship between decision
accuracy and decision speed [51]. AI may find itself to improve
medical decision accuracy by powering segmentation, classifica-
tion, and detection, and, at the same time, worsen it by increasing
radiologists’ workload to and beyond their optimal functioning
limits. In the light of these considerations, as direct recipients of
AI-based revolution in radiology, we should be aware that this
revolution can only take place when appropriate resources and
organizational support can be ensured to radiologists.
8.5 Conclusions
References
1. Weikert T, Winkel DJ, Bremerich J, Stieltjes B, Parmar V, Sauter
AW, Sommer G. Automated detection of pulmonary embolism in CT
pulmonary angiograms using an AI-powered algorithm. Eur Radiol.
2020;30(12):6545–53. https://doi.org/10.1007/s00330-020-06998-0.
2. Schmuelling L, Franzeck FC, Nickel CH, Mansella G, Bingisser R,
Schmidt N, et al. Deep learning-based automated detection of pul-
monary embolism on CT pulmonary angiograms: no significant effects
on report communication times and patient turnaround in the emergency
department nine months after technical implementation. Eur J Radiol.
2021;141:109816.
3. Olive-Gadea M, Crespo C, Granes C, Hernandez-Perez M, Pérez De La
Ossa N, Laredo C, Urra X, Carlos Soler J, Soler A, Puyalto P, Cuadras P,
Marti C, Ribo M. Deep learning based software to identify large vessel
occlusion on noncontrast computed tomography. Stroke. 2020;51:3133–
7. https://doi.org/10.1161/STROKEAHA.120.030326.
4. Seyam M, Weikert T, Sauter A, Brehm A, Psychogios MN, Blackham
KA. Utilization of artificial intelligence–based intracranial hemorrhage
detection on emergent noncontrast CT images in clinical workflow.
Radiology. 2022;4(2):1–6. https://doi.org/10.1148/ryai.210168.
5. Yang S, Yoon HI, Kim JS. Deep-learning-based automatic detection
and segmentation of brain metastases with small volume for stereotactic
ablative radiotherapy. Cancer. 2022;14:2555.
6. Turk O, Ozhan D, Acar E, Cetin T, Yilmaz M. Automatic detection of
brain tumors with the aid of ensemble deep learning architectures and
class activation map indicators by employing magnetic resonance images.
Z Med Phys. 2022; https://doi.org/10.1016/j.zemedi.2022.11.010.
7. Rahimpour M, Boellaard R, Jentjens S, Deckers W, Goffin K, Koole M.
A multi-label CNN model for the automatic detection and segmentation
of gliomas using [18 F] FET PET imaging. Eur J Nucl Med Mol Imaging.
2023; https://doi.org/10.1007/s00259-023-06193-5.
8. Akilandeswari A, Sungeetha D, Joseph C, Thaiyalnayaki K, Baskaran K,
Ramalingam RJ, Al-lohedan H, Al-dhayan DM, Karnan M, Hadish KM.
Automatic detection and segmentation of colorectal cancer with deep
residual convolutional neural network. Evid Based Complement Alternat
Med. 2022;2022:3415603.
9. Models P, Othman E, Mahmoud M, Dhahri H, Abdulkader H, Mahmood
A, Ibrahim M. Automatic detection of liver cancer using hybrid pre-
trained models. Sensors. 2022;22:5429.
10. Wang X, Li H, Zheng P. Automatic detection and segmentation of ovarian
cancer using a multitask model in pelvic CT images. Oxid Med Cell
Longev. 2022;2022:6009107.
8 Current Applications of AI in Medical Imaging 161