Basics of Image Processing
Basics of Image Processing
Ángel Alberich-Bayarri
Fuensanta Bellvís-Bataller Editors
Basics of
Image
Processing
The Facts and Challenges of
Data Harmonization to Improve
Radiomics Reproducibility
Imaging Informatics for Healthcare
Professionals
Series Editors
Peter M. A. van Ooijen, University Medical Center Groningen
University of Groningen, GRONINGEN, Groningen
The Netherlands
Erik R. Ranschaert, Department of Radiology, ETZ Hospital
Tilburg, The Netherlands
Annalisa Trianni, Department of Medical Physics, ASUIUD
UDINE, Udine, Italy
Michail E. Klontzas, Institute of Computer Science, Foundation
for Research and Technology (FORTH) & University Hospital
of Heraklion, Heraklion, Greece
The series Imaging Informatics for Healthcare Professionals is
the ideal starting point for physicians and residents and students
in radiology and nuclear medicine who wish to learn the basics in
different areas of medical imaging informatics. Each volume is a
short pocket-sized book that is designed for easy learning and ref-
erence.
The scope of the series is based on the Medical Imaging
Informatics subsections of the European Society of Radiology
(ESR) European Training Curriculum, as proposed by ESR and
the European Society of Medical Imaging Informatics (EuSoMII).
The series, which is endorsed by EuSoMII, will cover the curri-
cula for Undergraduate Radiological Education and for the level I
and II training programmes. The curriculum for the level III train-
ing programme will be covered at a later date. It will offer fre-
quent updates as and when new topics arise.
Ángel Alberich-Bayarri
Fuensanta Bellvís-Bataller
Editors
Basics of Image
Processing
The Facts and Challenges
of Data Harmonization
to Improve Radiomics
Reproducibility
Editors
Ángel Alberich-Bayarri Fuensanta Bellvís-Bataller
Founder and CEO VP of Clinical Studies
Quibim SL Quibim SL
Valencia, Spain Valencia, Spain
© EuSoMII 2023
This work is subject to copyright. All rights are solely and exclusively licensed by
the Publisher, whether the whole or part of the material is concerned, specifically
the rights of translation, reprinting, reuse of illustrations, recitation, broadcasting,
reproduction on microfilms or in any other physical way, and transmission or
information storage and retrieval, electronic adaptation, computer software, or by
similar or dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service
marks, etc. in this publication does not imply, even in the absence of a specific
statement, that such names are exempt from the relevant protective laws and
regulations and therefore free for general use.
The publisher, the authors, and the editors are safe to assume that the advice and
information in this book are believed to be true and accurate at the date of
publication. Neither the publisher nor the authors or the editors give a warranty,
expressed or implied, with respect to the material contained herein or for any
errors or omissions that may have been made. The publisher remains neutral with
regard to jurisdictional claims in published maps and institutional affiliations.
v
vi Preface
1 Era
of AI Quantitative Imaging�������������������������������������� 1
L. Marti-Bonmati and L. Cerdá-Alberich
2 Principles
of Image Formation in the Different
Modalities�������������������������������������������������������������������������� 27
P. A. García-Higueras and D. Jimena-Hermosilla
3 How
to Extract Radiomic Features from Imaging�������� 61
A. Jimenez-Pastor and G. Urbanos-García
4 Facts
and Needs to Improve Radiomics
Reproducibility ���������������������������������������������������������������� 79
P. M. A. van Ooijen, R. Cuocolo, and N. M. Sijtsema
5 Data
Harmonization to Address the Non-biological
Variances in Radiomic Studies���������������������������������������� 95
Y. Nan, X. Xing, and G. Yang
6 Harmonization
in the Image Domain ����������������������������117
F. Garcia-Castro and E. Ibor-Crespo
7 Harmonization
in the Features Domain ������������������������145
J. Lozano-Montoya and A. Jimenez-Pastor
vii
Era of AI Quantitative
Imaging 1
L. Marti-Bonmati
and L. Cerdá-Alberich
Fig. 1.1 The art of science and life representation based on phenotype imag-
ing for personalized classification and prediction of clinical outcomes to
achieve a diagnostic gain with respect to standard of care clinical practice
pared with that of the originating site. Metastases are usually bio-
logically different from the primary tumor. Therefore, imaging
has the potential to help targeting treatments if cancer hallmarks
and biological behavior are estimated from them.
Radiomics and imaging biomarkers are subrogated features
and parameters extracted from medical images, providing quanti-
tative information as the regional distribution and magnitude of
the evaluated property. They can also be clustered as nosological
signatures by the combination of relevant features into a single
value. This information is resolved in space, as parametric maps,
and time, through delta analysis of longitudinal changes.
Artificial intelligence (AI) offers a paradigm shift toward data-
driven tools decision-making that is revolutionizing medicine. AI
can be used to improve the process of data acquisition (such as
faster and higher quality MRI), extract new information from exist-
ing data (such as data labeling, lesion detection and segmentation,
and deep radiomics extraction for patients’ stratification), and gen-
erate prediction on future disease-related events (such as predictive
models on therapy response, and time-to-events for patient out-
comes). Nowadays, AI-powered imaging is being widely used in
cancer care, providing more reliable diagnosis and early detection,
improving screening results, adjusting follow-up schemes, aiding
in the discovery of new drugs, grading aggressiveness, defining best
treatments, and improving final prognostic outcomes (Fig. 1.2).
Fig. 1.2 Diagram of the AI and medical imaging innovation research path-
way, containing aspects related to the clinical question to answer, the data to
be employed, the model to be developed to predict a particular clinical out-
come and the proposed improvements for sustainability and reproducibility
of research
6 L. Marti-Bonmati and L. Cerdá-Alberich
Fig. 1.3 Schema of the AI-based workflow in medical imaging and oncol-
ogy, including tumor detection and segmentation, obtention of hallmarks in
terms of parametric maps, extraction of diagnosis models and tools for the
prediction of aggressiveness, overall survival, angiogenesis, cellularity, and
relationships between phenotyping–genotyping, for the development of a
Clinical Decision Support System (CDSS) that may impact decision on treat-
ment [based on the probability of treatment response, confidence level,
impact on radiotherapy (RT), etc.]. *DW diffusion-weighted, DCE dynamic
contrast-enhanced, MR magnetic resonance
Image Harmonization
One of the main challenges when developing AI models with
RWD is the large image heterogeneity caused by many different
vendors, scanners, protocols, acquisition parameters, and clinical
practice. One of the most promising areas of research regarding
image harmonization is the use of generative adversarial networks
(GANs) [14] to generate synthetic images which belong to a new
common framework space of standardized imaging data. GANs
make use of a generator and a discriminator to improve their abil-
ity of (1) creating new images (fake) as similar as possible to the
reference images (real) used as ground truth and (2) distinguish-
ing the real images from the fake ones. This method allows an
effective and efficient learning procedure where the only require-
ment is to have a well-defined ground truth; for instance, we may
target images in a particular imaging domain (e.g., images belong-
ing to a specific manufacturer, scanner, magnetic field strength or
type of weighting in MR images). Additionally, if paired images
are not available, the CycleGAN-based architecture is a great
alternative solution, which is a popular DL model used for image-
to-image translation tasks without paired examples. The models
are trained in an unsupervised manner using a collection of images
from the source and target domain that do not need to be related
in any way.
However, in the case of aiming to generate images in a new
common standardized imaging data space, we may not be able to
define the specific characteristics for these images to achieve a
better resolution and lower noise. Potential solutions may include
the use of the frequency space, which allows to isolate specific
components of the image, keeping its main information and
excluding the one related to its contrast, which is at the core of
image acquisition heterogeneity. This strategy can be used in
combination with autoencoders, which have demonstrated a good
10 L. Marti-Bonmati and L. Cerdá-Alberich
Image Segmentation
Volume of interest segmentation plays a crucial role in various
aspects of medical applications, such as quantifying the size and
shape of organs in population studies, detecting and extracting
lesions in disease analysis, defining computer-aided treatment
volumes, and surgical planning, among others. While manual seg-
mentation by medical experts was considered the ground truth, it
is expensive, time-consuming, and prone to disagreements among
readers. On the other hand, automatic segmentation methods offer
faster, cost-effective, and more reproducible results after manual
checking and editing [19].
Traditionally, segmentation relied on classical techniques like
region growing [20], deformable models [21], graph cuts [22],
clustering methods [23], and Bayesian approaches [24]. However,
in recent years, DL methods have surpassed these classical hand-
crafted techniques, achieving unprecedented performance in vari-
ous medical image segmentation tasks [25, 26]. Recent reviews
and advancements in DL for medical image segmentation are
available, focusing on improving network architecture, loss func-
tions, and training procedures [27]. Remarkably, it has been dem-
onstrated that standard DL models can be trained effectively using
limited labeled training images by making use of several transfer
learning techniques [28].
Although there is considerable variation in proposed network
architectures, they all share a common foundation: the use of con-
volution as the primary building block. Some alternative network
architectures have explored recurrent neural networks [29] and
attention mechanisms [30] but still rely on convolutional opera-
tions. However, recent studies suggest that a basic fully convolu-
tional network (FCN) with an encoder–decoder structure can
handle diverse segmentation tasks with comparable accuracy to
more complex architectures [31].
Convolutional neural networks (CNNs), including FCNs, owe
their effectiveness in modeling and analyzing images to key prop-
erties such as local connections, parameter sharing, and transla-
tion equivariance [32]. These properties provide CNNs with a
strong and valuable inductive bias, enabling them to excel in
1 Era of AI Quantitative Imaging 13
Fig. 1.4 Diagram of the AI image processing pipeline, including image har-
monization (spatial resolution, common framework, normalization), image
annotation and tumor extraction (3D segmentation), properties extraction
(parameters, deep features) and modeling, and personalized cancer phenotyp-
ing, prediction, and prognosis estimations
18 L. Marti-Bonmati and L. Cerdá-Alberich
References
1. Demicheli R, Fornili M, Querzoli P et al (2019) Microscopic tumor foci
in axillary lymph nodes may reveal the recurrence dynamics of breast
cancer. Cancer Commun 39:35. https://doi.org/10.1186/s40880-019-
0381-9
2. Cerdá Alberich L, Sangüesa Nebot C, Alberich-Bayarri A et al (2020) A
confidence habitats methodology in MR quantitative diffusion for the
classification of neuroblastic tumors. Cancers (Basel) 12(12):3858.
https://doi.org/10.3390/cancers12123858. PMID: 33371218; PMCID:
PMC7767170
3. Ni M, Zhou X, Lv Q et al (2019) Radiomics models for diagnosing
microvascular invasion in hepatocellular carcinoma: which model is the
best model? Cancer Imaging 19:60. https://doi.org/10.1186/s40644-019-
0249-x
4. Juan-Albarracín J, Fuster-Garcia E, Pérez-Girbés A et al (2018)
Glioblastoma: vascular habitats detected at preoperative dynamic
susceptibility-weighted contrast-enhanced perfusion MR imaging predict
survival. Radiology 287(3):944–954. https://doi.org/10.1148/
radiol.2017170845. Epub 2018 Jan 19. PMID: 29357274
5. Reeder SB, Yokoo T, França M et al (2023) Quantification of liver iron
overload with MRI: review and guidelines from the ESGAR and
SAR. Radiology 307(1):e221856. https://doi.org/10.1148/radiol.221856.
Epub 2023 Feb 21. PMID: 36809220; PMCID: PMC10068892
6. Martí-Aguado D, Jiménez-Pastor A, Alberich-Bayarri Á et al (2022)
Automated whole-liver MRI segmentation to assess steatosis and iron
quantification in chronic liver disease. Radiology 302(2):345–354.
https://doi.org/10.1148/radiol.2021211027. Epub 2021 Nov 16. PMID:
34783592
7. Kondylakis H, Kalokyri V, Sfakianakis S et al (2023) Data infrastructures
for AI in medical imaging: a report on the experiences of five EU proj-
ects. Eur Radiol Exp 7(1):20. https://doi.org/10.1186/s41747-023-
00336-x. PMID: 37150779; PMCID: PMC10164664
8. Marti-Bonmati L, Koh DM, Riklund K et al (2022) Considerations for
artificial intelligence clinical impact in oncologic imaging: an AI4HI
position paper. Insights Imaging 13(1):89. https://doi.org/10.1186/
s13244-022-01220-9. PMID: 35536446; PMCID: PMC9091068
22 L. Marti-Bonmati and L. Cerdá-Alberich
P. A. García-Higueras
and D. Jimena-Hermosilla
X-Ray Tube
The X-ray tube is the radiation source and is electrically powered
by the generator. It is composed of a protective housing and a
glass envelope in which a vacuum is created. Two main parts can
be found inside the glass envelope:
Generator
The generator is the device that transforms and accommodates the
power from the electrical grid to the needs of the X-ray tube. The
generator usually consists of two separate elements: the console
and the electric transformer. The operator can use the console to
define the radiological technique, which basically consists of
three adjustable parameters:
Photoelectric Effect
In photoelectric effect, the photon disappears being absorbed by
the atomic electrons that compose the medium. In other words,
the X-ray beam will lose a photon that will not reach the imaging
system. The photoelectric effect is a phenomenon that occurs
most likely at low X-ray beam energies (low kV).
Compton Effect
In the Compton effect, the photon interacts with an electron of the
medium and transfers part of its energy, the photon is scattered
after the interaction with a scattering angle. The process results in
a decrease of the photon energy and the emission of an atomic
electron.
Although the X-ray beam generation has not changed for many
decades and its interaction with matter is governed by invariant
physics laws, in the last decades image receptors have been evolv-
ing towards systems generically called digitals.
The way in which the image is obtained could be used to clas-
sify the X-ray equipment. A distinction could be made between
30 P. A. García-Higueras and D. Jimena-Hermosilla
Computed Tomography
Computed tomography (CT) was the first and earliest application
of digital radiology and is considered by many the greatest
advance in radiodiagnosis since the X-rays discovery [4]. In a
modern CT scan both the X-ray tube and the image detector are
rigidly mounted on a platform and rotate at the same time. This
structure is called gantry.
CT imaging systems are generally an array of solid-state detec-
tors fabricated by modules. The field of view (FOV) is delimited
by the physical extension of the detectors array. At the CT tube
exit, there are “shape filters” to adjust the intensity gradient of the
X-ray beam and a collimation system to limit the beam width
along the longitudinal axis. Another intrinsic element of a CT
scanner is the couch, which has precision motors for accurate
movements and lasers to centre the patients. The modes of CT
acquisitions can be classified into axial and helical (Fig. 2.1):
• Axial acquisition: The tube does not irradiate while the patient
moves between acquisition cycles.
• Helical acquisition: The tube has a spiral trajectory around the
patient in the form of helices. The ratio between the travel
length of the couch in a complete revolution and the slice
thickness is called pitch.
Reconstruction Algorithms
The basic principle of tomographic image reconstruction is the
accurate reproduction of the object from a set of projections taken
from different angles. Although tomographic reconstruction was
originally performed by algebraic methods, the slowness of the
calculation processes led to the use of analytical methods.
The most popular of the analytical methods is the filtered back-
projection (FBP) algorithm. This algorithm based on the Fourier
2 Principles of Image Formation in the Different Modalities 31
Spatial Resolution
Spatial resolution or high-contrast resolution is the ability of an
imaging system to discriminate and represent small details in an
image of an interest volume. This capability provides the details
of the anatomical structures of the tissues. The elements that
affect spatial resolution in a CT scan are summarized as follows:
Contrast Resolution
Contrast resolution or low-resolution contrast is the ability of an
imaging system to discriminate between different structures or
tissues within an interest volume. The elements that affect con-
trast resolution in a CT scan are:
Image Noise
Image noise refers to the presence of random signals or fluctua-
tions that are not related to the patient’s anatomy or pathology.
Noise has a negative impact on image quality, as it may hinder the
visualization of tissues or obscure subtle details. The most com-
mon sources of image noise in CT are as follows:
Artifacts
Artifacts are undesired structures resulting from distortions,
anomalies or interferences which appear in the image and do not
represent the patient’s anatomy or pathology. The most common
are artifacts from patient movements (intentional or uninten-
tional), artifacts from metallic objects in the volume of interest or
artifacts from hardware problems in the equipment.
Although the image quality characteristics describe different
aspects, they cannot be treated as completely independent factors
because the improvement of one of them is often obtained through
deterioration of one (or more) of the others.
2.2.1 Radiopharmaceuticals
Radioactive Decay
Each element from the periodic table has multiple isotopes. An
isotope of an element is a nucleus with the same atomic number
(number of protons) but differing from the number of neutrons.
Some isotopes are unstable and have some likelihood to undergo
a decay process. If so, they are called radionuclides. These radio-
nuclides may decay by different means.
PET is based on the decay path called beta plus decay (β+).
Radioisotopes with an excess of protons are likely to decay via β+
[8]. It means one of the protons from the nucleus is converted into
a neutron, emitting a positron, and an electronic neutrino in the
process. The difference of energy between the father radionuclide
and the daughter is shared between the positron and the neutrino.
A general expression for this process is expressed below, being X
the father nucleus, Y the daughter nucleus (losing one proton), e+
is the positron, and n e corresponds to antineutrino. Finally, Z and
A are the atomic and mass number, respectively,
A
Z X® Y + e + + ve
A
Z -1 (2.1)
2 Principles of Image Formation in the Different Modalities 37
Electron–Positron Annihilation
Immediately after the decay, the positron loses its kinetic energy
(10−1 cm) and, when it is almost at rest, interacts with an electron
from the tissue, resulting in the disappearance of both and the
emission of two photons of 511 kiloelectronvolts (keV) energy
moving in opposite directions. This interaction is called electron–
positron annihilation. Considering neither positron nor electron
have kinetic energy, the energy of the photons is a result of
Einstein’s energy-mass equation as follows, being me and mp the
masses of the electron and the positron, respectively, and c the
speed of light:
E = mc 2 = me c 2 + mp c 2 (2.2)
Therefore, the two opposed photons resulting from this pro-
cess are the basis of PET tomography. As these photons are very
energetic, they have high probability to escape from the body and
to be detected externally. Hence, placing two detectors in the line
the photons are moving and detecting them at the same time it is
known as the point where the annihilation (that is the decay) hap-
pened (Fig. 2.2).
A single detection (also called true detection) occurs when two
photons coming from an annihilation are detected in a short period
of time, called coincidence time (τ). At the same time, not only the
true detections occur, but other detections also occur deteriorating
the image quality and decreasing the quantitative information.
They are called random, scattered, and multiple coincidences
(Fig. 2.3).
Scattered Coincidences
Scattered coincidences are produced when a photon interacts
within the object studied via Compton scattering. These scattered
photons are registered by a detector out of the line of coincidence
of annihilation photons. Nonetheless, this effect may be reduced
using tungsten septa ring, absorbing photons at large angles, and
identifying and removing scattered photons, being the low energy
resolution of the detector the main drawback for this filtering
method.
38 P. A. García-Higueras and D. Jimena-Hermosilla
Random Coincidences
It could happen that two photons produced by different events are
detected by two opposed detectors within the timing coincidence
window, mistaken it for a true coincidence. The random events
rate probability (Crandom) is increased directly with the timing
coincidence window τ and the single event rate of each detector
(S1 and S2) as follows:
Crandom = 2t S1 S 2 (2.3)
As the activity increases so does the ratio between random/
single events rate probability. For this purpose, the use of septa
rings notably reduces this ratio besides the development of faster
detector with a lower timing coincidence window.
2 Principles of Image Formation in the Different Modalities 39
a b
c d
Fig. 2.3 Different events that result in a detection. True coincidences (a) are
important to create the image. Scattered, random and multiple coincidences
(b–d) worsen the image quality and it is necessary to identify and diminish
them. Figure adapted from published in [9]
Multiple Coincidences
There is a chance of two or more single events to be detected in
the timing coincidence window. In such case, the events are nor-
mally discarded because it turns out confusing for the event to be
positioned. Nevertheless, some annihilations contain information
of the quantity and spatial location of the decays. Therefore, under
specific circumstances, one of the lines is selected randomly from
the multiple detections.
40 P. A. García-Higueras and D. Jimena-Hermosilla
Radiofrequency Coils
They generate the radiofrequency radiation (RF) and are also
responsible for detecting the signal returned by the studied tis-
sues. The most important components are as follows:
Fig. 2.4 The particle rotation (spin) generates a magnetic moment μ. When
an external magnetic field B is applied, the magnetic moment tends to be
oriented in the direction of the magnetic field (B), forming an angle θ with it
and producing a precession motion at Larmor frequency. Figure adapted from
published in [13]
a b
Longitudinal Relaxation: T1
At the moment the RF pulse ends, the H nuclei release their
energy to the surrounding medium, so some of them oriented in
an antiparallel state return to their parallel state. A more homoge-
neous surrounding medium means a more coherent and uniform
energetic release.
50 P. A. García-Higueras and D. Jimena-Hermosilla
1, T2 or PD Weighted Images
T
All MRI have both T1 and T2 components. A correct selection of
TR and TE parameters allows a weighting of T1, T2 or a suitable
combination of both (PD weighted image). To summarize:
The magnetic gradient fields Gx, Gy, and Gz are activated to create
a spatial encoding along the three space directions. The Gz gradi-
ent is used to select the slice along the longitudinal axis, for the
transversal plane the Gy (phase encoding) and Gx (frequency
encoding) are used [14].
Phase encoding begins when Gy is activated. The rows that
receive a higher magnetic field precess at a higher frequency than
other rows that receive a lower magnetic field. When the Gy gradi-
ent closes there is a phase shift between the different rows whereby
the row in the plane can be uniquely identified.
The frequency encoding is activated by Gx, which is perpen-
dicular to Gy, so that each column will receive a different mag-
netic field. By Larmor’s Law the H nuclei of different columns
will precess at a different frequency. To prevent Gx and Gy from
overlapping making encoding of each row impossible, a bipolar
Gx gradient is applied, with two lobes of the same amplitude and
duration but in the opposite directions.
During the first lobe (−Gx) no signal is captured and it is used
to produce a phase shift that will be compensated for the one pro-
duced during the reading. When the second gradient (+Gx) is
applied, the echo signal is collected. The second lobe (+Gx) is
applied just after the first one and inverts the gradient over the
52 P. A. García-Higueras and D. Jimena-Hermosilla
-Space
K
The echo signal collected by the receiving antenna is subjected to
a series of stages before digitization [15]. Bandwidth (BW) is the
frequency range collected and accepted for digitization measured
in Hertz (Hz). The digitizing process of the echo signal is carried
out by measuring the voltage at regular time intervals called sam-
pling intervals (∆tm). The number of samples to be taken corre-
sponds to the number of pixels to be displayed in a row. Actually,
two components are generated in the process (real component and
imaginary component), although for simplicity of explanation, it
will be considered as a single component for the moment.
By Nyquist’s theorem, a signal can be mathematically recon-
structed if it is band-limited and the sampling rate is more than
two times the maximum frequency of the sampled signal.
The echo signal’s maximum frequency is found at the extreme
of the acquire band, that is, at BW/2. Applying Nyquist’s theorem,
the minimum reading frequency of the signal will be BW, and the
sampling intervals will be determined by
1 1
Dtm = = (2.5)
BW 2 · f max
Hence, for each encoding a line of digitized values of the echo
signal is obtained spaced by a sampling interval ∆tm. From the Gy
values employed, it is possible to arrange the lines obtained from
the different echoes as rows of a matrix in which the columns
would be separated by a time ∆tm and the rows by the time it takes
to transition from one echo signal to another, i.e., the TR. This
matrix, which constitutes the digitized data space in time-domain,
must be transformed to frequency-domain data to reconstruct the
image.
The signals that constitute the echo, obtained by the action of
the Gx gradient, belong to a frequency range that depend on the
position in the voxel plane, therefore, we obtain as many digitized
values expressed in spatial frequency scale (kx) as the number of
pixels to represent in a row.
54 P. A. García-Higueras and D. Jimena-Hermosilla
æ g ö æ FOVx ö
f max = ç ÷ · Gx · ç ÷ (2.6)
è 2p ø è 2 ø
Applying Nyquist’s theorem again, the spatial frequency is
given by
1 æ g ö
Dk x = =ç ÷ · Gx · Dtm (2.7)
FOVx è 2p ø
Since time t is taken based on the maximum value of the TE,
the values of kx are ordered on a line of spatial frequencies spaced
one interval apart ∆kx ordered symmetrically with respect to the
centre. This line constitutes a row in the k-space matrix.
In order to fill the matrix, it is necessary to collect as many
echoes as voxels in the column, which are encoded by the gradient
value Gy. Considering that the time of application of Gy is always
the same (ty), the difference between echoes will be the variation
of Gy gradient value, therefore:
1 æ g ö
Dk y = =ç ÷ · DG y · t y (2.8)
FOVy è 2p ø
This is how the data matrix that constitutes the K-space is
obtained, where each row is separated by ∆ky and each column by
∆kx, considering also that each matrix position (kx, ky) corresponds
to a value (signal strength).
Thus, the most external or peripheral line of the K-space matrix
will be filled with the highest value of Gy. Since high spatial fre-
quencies carry information about fast signal variations in space,
the most external rows of K-space carry information about the
spatial resolution of the image. Analogously, the central part of
the matrix is where the highest signal intensities are stored
because Gy has the lowest values of spatial frequencies, which
carry much information about contrast, i.e., the central part of the
K-space matrix carries information about contrast resolution.
2 Principles of Image Formation in the Different Modalities 55
Spatial Resolution
Parameters involved in spatial resolution are slice thickness, the
FOV or the image matrix size. To increase the spatial resolution,
the voxel volume has to be reduced by decreasing the slice thick-
ness, the FOV or by increasing the matrix size. These improve-
ments, as discussed above, result in a reduction of the SNR.
2 Principles of Image Formation in the Different Modalities 57
References
1. Brenner DJ, Hall EJ (2007) Computed tomography, an increasing source
of radiation exposure. New Engl J Med 357(22):2277–2284. https://doi.
org/10.1056/NEJMra072149
2. Johns HE, Cunningham JR (1983) The physics of radiology, 4th edn.
ISBN: 0-398-04669-7
3. Curry TS, Dowdey JE, Murry RC (1990) Christensen’s physics of diag-
nostic radiology, 4th edn. ISBN: 0-8121-1310-1
4. Herman GT (2009) Fundamentals of computerized tomography. Image
reconstruction from projections, 2nd edn. ISBN: 978-1-84628-737-7
5. Bracewell RN (2000) The Fourier Transform and its applications, 3rd
edn. ISBN: 0-07-303938-1
6. Aichinger H, Dierker J, Joite-BarfuB S, Säbel M (2012) Radiation expo-
sure and image quality in X-ray diagnostic radiology, 2nd edn. ISBN:
978-3-642-11240-9
7. Anand SS, Singh H, Dash AK (2009) Clinical applications of PET and
PET-CT. Med J Armed Forces India 65(4):353–358. https://doi.
org/10.1016/S0377-1237(09)80099-3
8. Shukla AK, Kumar U (2006) Positron emission tomography: an over-
view. J Med Phys 31(1):13–21. https://doi.org/10.4103/0971-6203.25665
9. Surti S (2015) Update on time-of-flight PET imaging. J Nucl Med
56(1):98–105. https://doi.org/10.2967/jnumed.114.145029
10. Cherry SR, Dahlbom M (2006) PET: physics, instrumentation and scan-
ners. ISBN: 978-0-387-32302-2
11. Cherry SR, Sorenson JA, Phelps ME (2012) Image quality in nuclear
medicine. Phys Nucl Med 233–251. https://doi.org/10.1016/b978-1-
4160-5198-5.00015-0
2 Principles of Image Formation in the Different Modalities 59
12. Weishaupt D, Köchli VD, Marincek B (2008) How does MRI work? An
introduction to the physics and function of magnetic resonance imaging,
2nd edn. ISBN: 978-3-540-30067-0
13. Harms SE, Morgan TJ, Yamanashi WS, Harle TS, Dodd GD (1984)
Principles of nuclear magnetic resonance imaging. Radiographics 4:26–
43. https://doi.org/10.1148/radiographics.4.1.26
14. Conolly S, Macovski A, Pauly J, Schenck J, Kwong KK, Chesler DA
(2000) The biomedical engineering handbook, 2nd edn. Magnetic
Resonance Imaging. ISBN: 0-8493-0461-X
15. Horowitz AL (1995) MRI Physics for Radiologists: a visual approach,
3rd edn. ISBN: 978-0-387-94372-5
16. Westbrook C, Kaut C (2006) MRI in clinical practice, 2nd edn. ISBN:
978-1-84628-161-7
How to Extract Radiomic
Features from Imaging 3
A. Jimenez-Pastor
and G. Urbanos-García
3.3.4 Standardization
DLRs have been used for disease diagnosis such as cancer type
prediction [26] or survival prediction [27]. These features can be
extracted through different DL architectures, in imaging, the most
common architectures are based on CNN.
The deep features extraction issue can be approached in differ-
ent ways and from different levels. On the one hand, the images
input can be at slice level, at volume level or at patient level. On
the other hand, DLRs can be extracted from either pretrained or
custom models.
Designing a model from scratch has the advantage of having a
network adjusted to the problem to solve. However, there may be
problems such as overfitting and class imbalance due to the lack
of available training datasets. To solve these problems, transfer
learning (TL) has been used as an alternative to construct new
models. TL is based on using a DL model pretrained with a natu-
ral image dataset and retrain the network with the desired medical
dataset to fine-tuning the hyperparameters. This approach has
been used in different studies. For example, in automatic polyp
detection in CT colonography [28], detection and classification of
breast cancer in microscope images [29] or pulmonary nodules in
thoracic CT images [30]. TL is usually applied using a pretrained
CNN model such as GoogleNet [31], Visual Geometry Group
Network (VGGNet) [32] or Residual Networks (ResNet) [33]
trained with the data from the ImageNet dataset. Deep features
from pretrained CNNs have achieved higher prediction accuracy
than hand-crafted radiomics signatures and clinical factors [34,
35]. However, TL has been arbitrarily configured in most studies,
and it is not evident whether good performance is obtained until
an evaluation of the model is performed.
DLRs can be extracted using both discriminative and genera-
tive deep learning networks [36]. Discriminative models are
supervised learning and use labels to distinguish classes, such as
distinguishing lesion from healthy tissue. Generative models are
unsupervised learning and extract general image features to gen-
erate new data with the same structure. In consequence, the
features extracted from generative models can be used as input to
a classifier.
3 How to Extract Radiomic Features from Imaging 73
v oting to obtain the best results [39, 40]. This voting can be soft,
hard or adaptative. Feature-level fusion consists of concatenating
the HCRs and DLRs vectors, applying features reduction to avoid
overfitting and using them as input to a model, obtaining better
results in lung cancer survival models [27] or tumor detection
[38]. Thus, radiomics and deep features have proven to be two
novel technologies that have a high potential for early detection,
prediction of treatment response and prognosis of the disease.
Figure 3.2 shows the different approaches introduced along the
chapter, going from HCR to the different approaches to extract
DLR and how they can be combined.
References
1. Chetan MR, Gleeson FV (2021) Radiomics in predicting treatment
response in non-small-cell lung cancer: current status, challenges and
future perspectives. Eur Radiol 31(2):1049–1058. https://doi.org/10.1007/
s00330-020-07141-9. Epub 2020 Aug 18. PMID: 32809167; PMCID:
PMC7813733
2. Du P, Liu X, Shen L, Wu X, Chen J, Chen L, Cao A, Geng D (2023)
Prediction of treatment response in patients with brain metastasis receiv-
ing stereotactic radiosurgery based on pre-treatment multimodal MRI
radiomics and clinical risk factors: A machine learning model. Front
Oncol 13:1114194. https://doi.org/10.3389/fonc.2023.1114194. PMID:
36994193; PMCID: PMC10040663
3. Huynh LM, Hwang Y, Taylor O, Baine MJ (2023) The use of MRI-
derived radiomic models in prostate cancer risk stratification: a critical
review of contemporary literature. Diagnostics (Basel) 13(6):1128.
13(6):1128. https://doi.org/10.3390/diagnostics13061128. PMID:
36980436; PMCID: PMC10047271
4. Zhang Y, Yang Y, Ning G, Wu X, Yang G, Li Y (2023) Contrast computed
tomography-based radiomics is correlation with COG risk stratification
of neuroblastoma. Abdom Radiol (NY) https://doi.org/10.1007/s00261-
023-03875-4. Epub ahead of print. PMID: 36951989
5. Cui Y, Li Z, Xiang M, Han D, Yin Y, Ma C (2022) Machine learning mod-
els predict overall survival and progression free survival of non-surgical
esophageal cancer patients with chemoradiotherapy based on CT image
radiomics signatures. Radiat Oncol. 17(1):212. https://doi.org/10.1186/
s13014-022-02186-0. PMID: 36575480; PMCID: PMC9795769
6. Chu F, Liu Y, Liu Q, Li W, Jia Z, Wang C, Wang Z, Lu S, Li P, Zhang Y,
Liao Y, Xu M, Yao X, Wang S, Liu C, Zhang H, Wang S, Yan X, Kamel
IR, Sun H, Yang G, Zhang Y, Qu J (2022) Development and validation of
MRI-based radiomics signatures models for prediction of disease-free
survival and overall survival in patients with esophageal squamous cell
carcinoma. Eur Radiol. 32(9):5930–5942. https://doi.org/10.1007/
s00330-022-08776-6. Epub 2022 Apr 6. PMID: 35384460
7. Chen W, Qiao X, Yin S, Zhang X, Xu X (2022) Integrating Radiomics
with Genomics for Non-Small Cell Lung Cancer Survival Analysis. J
Oncol. 2022:5131170. https://doi.org/10.1155/2022/5131170. PMID:
36065309; PMCID: PMC9440821
8. Feng Q, Ding Z (2020) MRI Radiomics Classification and Prediction in
Alzheimer’s Disease and Mild Cognitive Impairment: A Review. Curr
Alzheimer Res. 17(3):297–309. https://doi.org/10.2174/1567205017666
200303105016. PMID: 32124697
9. Pujadas ER, Raisi-Estabragh Z, Szabo L, McCracken C, Morcillo CI,
Campello VM, Martín-Isla C, Atehortua AM, Vago H, Merkely B, Mau-
rovich-Horvat P, Harvey NC, Neubauer S, Petersen SE, Lekadir K 2022
76 A. Jimenez-Pastor and G. Urbanos-García
28. Summers RM, Johnson CD, Pusanik LM, Malley JD, Youssef, AM, Reed
JE (2001) Automated polyp detection at CT colonography: feasibility
assessment in a human population. Radiology 219(1):51–59
29. Wang Y, Sun L, Ma K, Fang J (2018) Breast cancer microscope image
classification based on CNN with image deformation. In Image Analysis
and Recognition: 15th International Conference, ICIAR 2018, Póvoa de
Varzim, Portugal 27–29;2018. Proceedings 15 (pp. 845–852). Springer
International Publishing
30. Dehmeshki J, Amin H, Valdivieso M, Ye X (2008) Segmentation of pul-
monary nodules in thoracic CT scans: a region growing approach. IEEE
transactions on medical imaging 27(4):467–480.
31. Szegedy C et al (2015) Going deeper with convolutions. En Proceedings
of the IEEE conference on computer vision and pattern recognition.
pp 1–9
32. Simonyan K, Zisserman A (2014) Very deep convolutional networks for
large-scale image recognition. arXiv preprint arXiv:1409.1556
33. He K et al (2016) Deep residual learning for image recognition. In:
Proceedings of the IEEE conference on computer vision and pattern rec-
ognition. pp 770–778
34. Zhu Y et al (2019) A deep learning radiomics model for preoperative
grading in meningioma. Eur J Radiol 116:128–134
35. Zheng X et al (2020) Deep learning radiomics can predict axillary lymph
node status in early-stage breast cancer. Nat Commun 11:1–9
36. Afshar P et al (2019) From handcrafted to deep-learning-based cancer
radiomics: challenges and opportunities. IEEE Signal Process Mag
36(4):132–160
37. Echaniz O, Graña M (2017) Ongoing work on deep learning for lung
cancer prediction. In: Biomedical applications based on natural and arti-
ficial computing: international work-conference on the interplay between
natural and artificial computation, IWINAC 2017, Corunna, Spain, June
19–23, 2017, Proceedings, Part II. Springer International Publishing,
pp 42–48
38. Fu L et al (2017) Automatic detection of lung nodules: false positive
reduction using convolution neural networks and handcrafted features.
In: Medical imaging 2017: computer-aided diagnosis. SPIE, pp 60–67
39. Hassan AH, Wahed ME, Metwally MS, Atiea MA (2022) A hybrid
approach for classification breast cancer histopathology images. Frontiers
in scientific research and technology 3(1):1–10
40. Liu S et al (2017) Pulmonary nodule classification in lung cancer screen-
ing with three-dimensional convolutional neural networks. J Med
Imaging 4(4):041308
Facts and Needs to Improve
Radiomics Reproducibility 4
P. M. A. van Ooijen , R. Cuocolo ,
and N. M. Sijtsema
4.1 Introduction
Fig. 4.1 Example of quantitative imaging. Lung volumes and areas of low
attenuation are provided in cm3 and in percentage of low attenuation
Table 4.1 Overview of the steps in the radiomics process, the related factors
influencing the accuracy and reproducibility, and the possible solutions
reported in literature
Factors influencing
Step accuracy Possible solutions
Acquisition Digital image Protocol standardisation/
pre-processing harmonization
Voxel size/slice Histogram normalization
thickness Interpolation
Reconstruction filters
Contrast enhancement
protocol
Image noise
Patient size
Artifacts
Segmentation Lack of accuracy Fixed protocols
Inter-reader variability Consensus segmentation
Intra-reader variability
Validation
Feature Feature definition IBSI guidelines + digital
extraction Feature parameter phantom
setting Well accepted, open source,
Feature implementation feature implementations
Software used for Delta-radiomics
feature extraction
Model Feature selection IBSI guidelines
construction Machine learning Inter-software comparison
model selection
Cut-off selection
Model validation
4.2.1 Acquisition
kept the same, the images could still result in different radiomic
features values.
Another problem is the variation in the reconstruction param-
eters as defined by local protocols. These reconstruction parame-
ters have shown their influence to the appearance of the imaging
data to such an extent that they affect quantitative measurements
and radiomic features. Examples of such reconstruction parame-
ters are properties like the in-plane resolution, the slice thickness
and applied reconstruction kernels. Although previous imaging
studies have shown the effects of slice thickness and reconstruc-
tion kernels on computed features, between ~5% and ~25% of
radiomics studies prior to 2020 did not even report their study
imaging protocols. Most of those who did report their imaging
protocols only included the slice thickness information [2].
Additional to the scan protocol, the contrast enhancement pro-
tocol also plays a major role in the presentation of the image. This
includes the injection protocol itself (bolus timing and size) but
also the type of contrast media used (e.g., its iodine concentration
in CT scanning). One must keep in mind that the effect of the
contrast media can extend beyond the targeted area. For example,
with intravenous injection of contrast media into the blood for the
enhancement of the arteries this enhancement can also be appar-
ent beyond the wall of the arteries because of partial volume
effects. For example, Kristanto et al. showed a strong positive cor-
relation between lumen contrast enhancement and mean plaque
HU-value [6].
Scan artifacts also can hamper the determination of quantita-
tive features. These include not only artifacts caused by alien
objects such as metal implants (e.g., pacemakers, dental fillings,
hip/knee prostheses), but also those caused by inaccurate acquisi-
tion (e.g., incorrect triggering/gating or contrast timing) or volun-
tary and involuntary movement of the patient.
Finally, the patients themself also play a role in the determina-
tion of radiomic features. Different size patients or female patients
with different size breasts can—because of the disturbance of
fatty tissue—have different quantitative measures for the same
structure.
84 P. M. A. van Ooijen et al.
4.2.2 Segmentation
References
1. Steiger P, Sood R (2019) How can radiomics be consistently applied
across imagers and institutes. Radiology 291:60–61
2. Zhao B (2021) Understanding sources of variation to improve the repro-
ducibility of radiomics. Front Oncol 11:633176
3. Lennartz S, O’Shea A, Parakh A, Persigehl T, Baessler B, Kambadakone
A (2022) Robustness of dual-energy CT-derived radiomic features across
three different scanner types. Eur Radiol 32:1959–1970
4. Pfaehler E, Zhovannik I, Wei L, Boellaard R, Dekker A, Monshouwer R,
El Naqa I, Bussink J, Gillies R, Wee L, Traverso A (2021) A systematic
review and quality of reporting checklist for repeatability and reproduc-
ibility of radiomic features. Phys Imaging Radiat Oncol 20:69–75
5. Traverso A, Wee L, Dekker A, Gillies R (2018) Repeatability and repro-
ducibility of radiomic features: a systematic review. Int J Radiat Oncol
Biol Phys 102(4):1143–1158
6. Kristanto W, van Ooijen PMA, Greuter MJW, Groen JM, Vliegenthart R,
Oudkerk M (2013) Non-calcified coronary atherosclerotic plaque visual-
ization on CT: effects of contrast-enhancement and lipid-content frac-
tions. Int J Cardiovasc Imaging 29:1137–1148
92 P. M. A. van Ooijen et al.
netic field variations in MRI, are some of the factors that cause
differences in these data. Such variations can have serious
implications for the reproducibility of radiomic features, even
when a fixed acquisition protocol is used for different brands
of scanners.
For example, researchers have investigated the reproduc-
ibility of radiomic features on different scanners despite using
the same acquisition protocol and found substantial differ-
ences [2]. The reproducibility of radiomic features ranged
from 16% to 85%, indicating that even with a fixed protocol,
there are still significant differences in the images produced by
different scanners. Similarly, Sunderland et al. [3] found a
large variation in the standard uptake value (SUV) of different
brands of scanners. They discovered that newer scanners had a
much higher maximum SUV compared to older ones, indicat-
ing that the heterogeneity of acquisition devices could signifi-
cantly impact the interpretation of imaging data.
These findings suggest that the variability in acquisition
devices can significantly affect the reproducibility and reliabil-
ity of imaging data, which can have serious implications for
clinical decision-making. To address this issue, there is a need
for data harmonization strategies to ensure that imaging data is
98 Y. Nan et al.
Fig. 5.3 Two typical ways (featurewise and samplewise) of automatic data
harmonization
104 Y. Nan et al.
Fig. 5.4 Harmonization approaches. Blue blocks represent methods that can
be used for samplewise harmonization only, while the orange blocks corre-
spond to methods that can be used for two harmonization schemes. The yel-
low block indicates invariant representation learning approach which is
mainly used to develop harmonized models
x - min ( x )
x¢ = , (5.2)
max ( x ) - min ( x )
respectively.
In addition to normalization/standardization, the ComBat
algorithm, as described in [15, 16], was proposed for featurewise
harmonization. For instance, researchers used ComBat to harmo-
nize the image-derived features from multicentre MRI datasets
[16]. It utilized empirical Bayes shrinkage to accurately estimate
the mean and variance for each batch of data. These estimates
were then used to harmonize the data across cohorts. The first step
is standardizing the data to ensure similar overall mean and vari-
ance, followed by the empirical Bayes estimation with parametric
empirical priors. The resulting adjusted bias estimators were then
used in the location-scale model-based functions to harmonize the
data.
Another type of location and scale methods are based on the
alignment of data distributions, using cumulative distribution
functions or pdfs. For instance, Wrobel et al. [17] proposed a
method to harmonize MRI multicentre data, which aligns the
voxel intensities of the source dataset with the target cumulative
distribution functions by estimating a non-linear intensity trans-
formation. In another study [18], the empirical density was esti-
mated and the distance between probability density functions was
calculated. Common features from different datasets were selected
first, and then their probability density functions were estimated
to determine the most suitable matching offsets. The harmonized
data was obtained by subtracting the estimated offsets from the
source cohorts.
Clustering Methods
Clustering methods are commonly used in data harmonization to
group data samples based on their distances. In clustering, dis-
tance measures the similarity or dissimilarity between pairs of
data samples. The distance between two observations is typically
calculated based on the values of their attributes or features. The
aim of clustering is to create subsets or clusters of samples that are
5 Data Harmonization to Address the Non-biological Variances… 107
more like each other than to those in other clusters. This grouping
can help to harmonize the data by creating a more uniform repre-
sentation of the samples that can be used for subsequent analysis.
Figure 5.5 illustrates the steps when using clustering methods for
data harmonization.
Several factors, such as the choice of clustering algorithm, the
selection of distance metrics, the pre-processing of data, and the
determination of the number of clusters, can influence the quality
of harmonization obtained through clustering methods.
Clustering algorithm: The selection of clustering algorithm
can affect the quality of harmonization. Different algorithms have
different assumptions and properties and may perform different
on various types of data. For instance, k-means assumes spherical
clusters and is sensitive to initialization, while hierarchical clus-
tering can handle non-spherical clusters but requires more compu-
tational costs.
Distance metrics: The choice of distance metric can also
impact the quality of harmonization. Different distance metrics
lead to different clustering results, as the similarity or dissimilar-
ity between samples is calculated in different ways.
108 Y. Nan et al.
Matching Methods
Matching methods in data harmonization are used to align data
collected from different sources that may have different formats
or structures. Resampling is the common matching method used
for automatic data harmonization. Resampling, also known as
resizing, is a method that involves altering the dimensions or reso-
lution of images or signals to match those of other datasets. This
method can be used to harmonize data collected from different
sources with varying resolutions or image sizes. In radiomic
study, the reproducibility of radiomic features is heavily affected
by the voxel/pixel size (refers to the physical length of a single
pixel in the CT/MRI image).
Synthesis Methods
Synthesis is a method used to generate samples that belong to a
specific modality or domain, effectively harmonizing multi-cohort
datasets. This approach simplifies the task of data harmonization
by considering each cohort as a distinct style and transferring all
samples to a common style. Synthesis techniques can be divided
into paired synthesis and unpaired synthesis, depending on the
features of the training sample. Paired synthesis is used when cor-
responding samples from different cohorts are available, while
unpaired synthesis is used when such correspondence is absent (in
Fig. 5.6).
Paired synthesis approaches are trained on paired samples that
originate from the same object but are obtained using different
protocols (e.g., CT scans collected from same patient with differ-
ent scanners). These techniques are developed to learn how to
transform data between the source and reference cohorts. For
example, Park et al. proposed “deep harmonics” for CT slice
techniques can provide insight into the possible reasons for incon-
sistent data representations that contribute to bias in data-based
models. By analysing these insights, researchers can determine
whether the biasing artifacts are due to inadequate data harmoni-
zation before the learning phase. Additionally, local explanatory
methods can identify out-of-distribution examples that may relate
to data harmonization issues, such as equipment miscalibration or
changes in data capture protocols. Improved data harmonization
can benefit XAI by standardizing all data and eliminating cohort
biases [31, 32]. In summary, we anticipate an exciting cross-
disciplinary research area at the intersection of harmonization and
XAI.
References
1. Nan Y et al (2022) Data harmonization for information fusion in digital
healthcare: a state-of-the-art systematic review, meta-analysis and future
research directions. Inf Fusion 82:99
2. Berenguer R et al (2018) Radiomics of CT features may be nonreproduc-
ible and redundant: influence of CT acquisition parameters. Radiology
288(2):407–415
3. Sunderland JJ, Christian PE (2015) Quantitative PET/CT scanner perfor-
mance characterization based upon the society of nuclear medicine and
molecular imaging clinical trials network oncology clinical simulator
phantom. J Nucl Med 56(1):145–152
4. Yamashita R et al (2020) Radiomic feature reproducibility in contrast-
enhanced CT of the pancreas is affected by variabilities in scan parame-
ters and manual segmentation. Eur Radiol 30(1):195–205
5. Jha A et al (2021) Repeatability and reproducibility study of radiomic
features on a phantom and human cohort. Sci Rep 11(1):1–12
6. Emaminejad N, Wahi-Anwar MW, Kim GHJ, Hsu W, Brown M, McNitt-
Gray M (2021) Reproducibility of lung nodule radiomic features: multi-
variable and univariable investigations that account for interactions
between CT acquisition and reconstruction parameters. Med Phys
48:2906
7. Kim M, Jung SC, Park JE, Park SY, Lee H, Choi KM (2021)
Reproducibility of radiomic features in SENSE and compressed SENSE:
impact of acceleration factors. Eur Radiol 31:1–14
8. Saeedi E et al (2019) Radiomic feature robustness and reproducibility in
quantitative bone radiography: a study on radiologic parameter changes.
J Clin Densitom 22(2):203–213
114 Y. Nan et al.
F. Garcia-Castro (*)
Department of Technology and Innovation, Quibim, Valencia, Spain
e-mail: [email protected]
E. Ibor-Crespo
Department of AI Research, Quibim, Valencia, Spain
1
Since we will be discussing mainly tomographic image, we will refer to
voxels instead of pixels.
6 Harmonization in the Image Domain 119
models. If when acquiring a lung CT, the patient does not perform
the full inspiration cycle properly, the study will not be adequate
to perform any kind of volumetric analysis [10]. In multiparametric
prostate MRI (mpMRI), if the patient underwent suboptimal rec-
tal preparation, the rectum could show an excessive amount of
gas, creating susceptibility artifacts that especially affect the dif-
fusion weighted imaging (DWI) sequence with a large degree of
unwanted deformation [11], affecting image registration or organ
segmentation. Or if during the acquisition of a FDG PET/CT (flu-
orodeoxyglucose positron emission tomography + computed
tomography) scan the patient is not placed with the arms raised
above the head but alongside the body, beam hardening artifacts
and field-of-view (FOV) truncation artifacts might appear [12].
Poor patient preparation, in general, cannot be solved by har-
monization techniques. Standardization of patient preparation for
each type of imaging study is the best way to ensure reducing
variability in the cases where this procedure is needed. However,
there is no consensus patient preparation protocols depicted for all
reporting guidelines. For instance, the PI-RADS v2.1 prostate
mpMRI reading guidelines of the American College of Radiology
(ACR) specifically states that there is no consensus patient prepa-
ration that demonstrates an improvement in diagnostic accuracy.
Nevertheless, artifacts introduced by lack of patient preparation
will significantly hinder the performance of AI models and tradi-
tional computer vision algorithms, as anatomy can suffer heavy
deformations or other unwanted modifications.
While IHTs may not be the ideal solution for addressing this
kind of issues, they are particularly useful when it comes to deal-
ing with variability sources that directly impact the signal inten-
sity and contrast of an image. Factors such as the vendor of the
equipment, the specific scanner model, the firmware version, the
reconstruction algorithms employed, and the acquisition proto-
cols utilized can all affect image contrast in different ways
depending on the modality.
For instance, the tube voltage in CT scans and the repetition
time (TR) in MRI scans can fall within acceptable ranges for diag-
nostic purposes. However, the differences in SNR and image con-
trast that they introduce can hinder the performance of AI
122 F. Garcia-Castro and E. Ibor-Crespo
Intensity Scaling
Intensity scaling is a technique used in image processing to adjust
the contrast and brightness of an image by scaling the range of
voxel intensity values. In other words, it involves mapping the
original intensity values of an image to a new range of values.
The scaling process of a grayscale digital image typically
involves two steps: normalization and rescaling. In the normaliza-
tion step, the minimum and maximum intensity values in the
image are identified. Then, the intensity values of all the voxels in
the image are shifted and scaled to be in the range of [0,1] using
Eq. 6.1,
I i - I min
Ni = (6.1)
I max - I min
where Ii is an image voxel, Imin and Imax are the image minimum
and maximum intensity, respectively, and Ni is the normalized
voxel. After normalization, the image is rescaled to a new range of
intensity values. This is usually done to enhance the contrast of
the image by stretching the intensity range to occupy the full
available range of values. The new intensity values are obtained
using Eq. 6.2,
-Score Normalization
Z
Z-score is a statistical measure that is used to evaluate how many
standard deviations a data point is from the mean of a dataset. In
the context of image processing, Z-score normalization is a tech-
nique that is used to normalize the intensity of voxels in an image.
It is a linear transformation method that scales the voxel values to
have a mean of zero and a standard deviation of one.
Z-score normalization is applied to images by computing the
mean and standard deviation of the intensity values of all the vox-
els in the image. The mean is subtracted from each voxel value,
and the result is divided by the standard deviation (Eq. 6.3). This
transforms the voxel values such that the mean of the image is
zero and the standard deviation is one.
Ii - m
Zi = (6.3)
s
where Zi is a Z-score normalized voxel and μ and σ are the mean
and standard deviation of all the image voxels, respectively.
As a standalone method for image harmonization, the Z-score
might not be applicable in all scenarios, as it assumes that the
distribution of voxel intensities in an image is Gaussian. If the
distribution is non-Gaussian, the normalization may not be appro-
priate. Additionally, Z-score normalization can only adjust the
overall brightness and contrast of an image and may not be able to
correct for more complex artifacts or variability in image acquisi-
tion. However, it has been used in recent research in different situ-
ations, such as a normalization method in MRI of head and neck
cancer [18] or as part of normalization strategies for radiomic
pipelines [19]. It can also be of great help as part of the prepro-
cessing pipeline of an AI model training, as data with an average
close to zero can help speed up convergence in specific scenarios
[20].
Histogram Equalization
Histogram equalization (HE) is a technique used in image pro-
cessing to enhance the contrast of an image by redistributing
voxel values in the image’s histogram. It works by increasing the
128 F. Garcia-Castro and E. Ibor-Crespo
global contrast of the image, which can reveal hidden details and
improve the overall quality of the image. While HE does not guar-
antee obtaining the same contrast across a dataset, it will create a
similar effect on all of them due to the flattening of the histogram.
HE works by transforming the original image’s voxel intensi-
ties to a new set of intensities such that the cumulative distribution
function (CDF) of the resulting image is as flat as possible. The
CDF is a measure of the distribution of voxel intensities in the
image.
The histogram equalization algorithm is a two-step process.
The first step is to calculate the histogram of the input image,
which is a plot of the frequency of occurrence of each gray level
in the image. The second step is to calculate the cumulative distri-
bution function of the histogram, which represents the number of
voxels with intensity levels less than or equal to a given level. The
image is then transformed by mapping the original voxel intensi-
ties to their new values in a way that equalizes the CDF as seen in
Eq. 6.4.
æ ( L - 1) ö
Ei = round ç * CDF ( I i ) ÷ (6.4)
è M ø
where Ei is the new voxel intensity value, Ii is the original voxel
intensity value, M is the total number of voxels in the image and
L is the number of possible voxel intensity levels. Figure 6.2
shows the effect of HE on a T2w prostate MRI slice.
Histogram equalization may not work well for images with a
bimodal or multimodal histogram, where there are several peaks
in the histogram. In such cases, adaptive histogram equalization
techniques such as contrast limited adaptive histogram equaliza-
tion (CLAHE) may be used [21].
CLAHE is a modified version of the traditional HE technique
which overcomes the limitations of HE by dividing the image into
small rectangular regions called tiles, and then applying the HE
technique to each tile separately. The size of the tiles is usually
chosen based on the size of the features of interest in the image.
For example, for medical images such as MRI scans, smaller tiles
can be used to capture the fine details of the image.
6 Harmonization in the Image Domain 129
Fig. 6.2 Effect of HE on a T2w prostate MRI slice. Top left, original slice.
Top right, equalized slice. Bottom left, original histogram. Bottom right,
equalized histogram
Histogram Matching
Histogram matching, also known as histogram specification, is a
technique used to match the histogram of one image to another,
typically a reference image. It has been applied as a normalization
technique of medical images for many years [23]. The goal of
130 F. Garcia-Castro and E. Ibor-Crespo
Fig. 6.3 Effect of histogram matching on a FLAIR MRI slice. Top left, input
FLAIR image. Top middle, reference FLAIR image. Top right, result image
after histogram matching. Bottom left, input histogram and CDF. Bottom
middle, reference histogram and CDF. Bottom right, result histogram and
CDF after histogram matching. Note that the result image CDF matches the
shape of the reference image CDF
6 Harmonization in the Image Domain 131
Autoencoders
Autoencoders are a type of CNN that can be trained to learn a
compressed representation of input images by encoding them into
a low-dimensional latent space. The encoded information is then
formatted to solve a particular task, which depends on the archi-
tecture design and loss function, among others.
In medical imaging, autoencoders have been used for various
tasks, such as denoising [26] or segmentation [27]. Recently,
autoencoders have also been explored for image harmonization
purposes [28]. The basic idea behind image harmonization using
autoencoders is to train the network to learn the underlying fea-
tures of a set of medical images and then use this knowledge to
generate new images that have similar features, but with reduced
variability in appearance.
To achieve image harmonization using autoencoders, the net-
work is first trained using a set of input medical images. The
encoder part of the network is used to extract features from the
images, which are then compressed into a lower-dimensional
latent space. The decoder part of the network then takes this com-
pressed representation and reconstructs an output image that is as
close as possible to the original input image.
Once the autoencoder has been trained, it can be used to gener-
ate new images that are similar to the original input images but
with reduced variability. To do this, a new image is first fed
through the encoder to generate its latent representation. This
latent representation is then fed into the decoder to generate a new
image that is similar in appearance to the original input image but
has been harmonized to match the features of the training set.
Figure 6.4 shows a generic diagram of an autoencoder architec-
ture.
There are some limitations that should be considered. One
limitation is that the quality of the harmonized image is heavily
dependent on the quality and quantity of the training data. If the
134 F. Garcia-Castro and E. Ibor-Crespo
6.4 Conclusions
References
1. Smith TB, Zhang S, Erkanli A, Frush D, Samei E (2021) Variability in
image quality and radiation dose within and across 97 medical facilities.
J Med Imaging (Bellingham) 8(5):052105. https://doi.org/10.1117/1.
JMI.8.5.052105. Epub 2021 May 8. PMID: 33977114; PMCID:
PMC8105613
2. Smith NB, Webb A (2010) Introduction to medical imaging: physics,
engineering and clinical applications. Cambridge University Press
3. Yan W, Huang L, Xia L, Gu S, Yan F, Wang Y, Tao Q (2020) MRI manu-
facturer shift and adaptation: increasing the generalizability of deep
learning segmentation for MR images acquired with different scanners.
Radiol Artif Intell 2(4):e190195. https://doi.org/10.1148/
ryai.2020190195. PMID: 33937833; PMCID: PMC8082399
4. Shukla-Dave A, Obuchowski NA, Chenevert TL, Jambawalikar S,
Schwartz LH, Malyarenko D, Huang W, Noworolski SM, Young RJ,
Shiroishi MS, Kim H, Coolens C, Laue H, Chung C, Rosen M, Boss M,
Jackson EF (2019) Quantitative imaging biomarkers alliance (QIBA) rec-
ommendations for improved precision of DWI and DCE-MRI derived
biomarkers in multicenter oncology trials. J Magn Reson Imaging
49(7):e101–e121. https://doi.org/10.1002/jmri.26518
5. Schellinger PD, Jansen O, Fiebach JB, Hacke W, Sartor K (1999) A stan-
dardized MRI stroke protocol: comparison with CT in hyperacute intra-
140 F. Garcia-Castro and E. Ibor-Crespo
Morris JC, Albert MS, Grabe HJ, Resnick SM, Bryan NR, Wittfeld K,
Bülow R, Wolk DA, Shou H, Nasrallah IM, Davatzikos C, iSTAGING
and PHENOM Consortia (2022) Deep generative medical image harmo-
nization for improving cross-site generalization in deep learning predic-
tors. J Magn Reson Imaging 55(3):908–916. https://doi.org/10.1002/
jmri.27908. Epub 2021 Sep 25. PMID: 34564904; PMCID: PMC8844038
31. Zhu JY, Park T, Isola P, Efros AA (2017) Unpaired image-to-image trans-
lation using cycle-consistent adversarial networks. In: IEEE international
conference on computer vision (ICCV), pp 2242–2251
32. Mirza M, Osindero S (2014) Conditional generative adversarial nets.
arXiv preprint arXiv:1411.1784
33. Chen J, Wee L, Dekker A, Bermejo I (2022) Improving reproducibility
and performance of radiomics in low-dose CT using cycle GANs. J Appl
Clin Med Phys 23(10):e13739. https://doi.org/10.1002/acm2.13739.
Epub 2022 Jul 30. PMID: 35906893; PMCID: PMC9588275
34. Fatania K, Clark A, Frood R, Scarsbrook A, Al-Qaisieh B, Currie S, Nix
M (2022) Harmonisation of scanner-dependent contrast variations in
magnetic resonance imaging for radiation oncology, using style-blind
auto-encoders. Phys Imaging Radiat Oncol 22:115–122. https://doi.
org/10.1016/j.phro.2022.05.005. PMID: 35619643; PMCID:
PMC9127401
35. Dewey BE, Zhao C, Reinhold JC, Carass A, Fitzgerald KC, Sotirchos ES,
Saidha S, Oh J, Pham DL, Calabresi PA, van Zijl PCM, Prince JL (2019)
DeepHarmony: a deep learning approach to contrast harmonization
across scanner changes. Magn Reson Imaging 64:160–170. https://doi.
org/10.1016/j.mri.2019.05.041. Epub 2019 Jul 10. PMID: 31301354;
PMCID: PMC6874910
Harmonization
in the Features Domain 7
J. Lozano-Montoya
and A. Jimenez-Pastor
7.1 Introduction
CT Scans
One of the most studied factors influencing reconstruction in CT
scans is the voxel size. However, some studies also investigate the
impact of image discretization on radiomic features [1]. When
different acquisition modes and image reconstructions were
applied to CT, most features were found to be redundant, with
only 30% of them being reproducible across test-retest, with a
concordance correlation coefficient (CCC) of at least 0.90 [7].
When a phantom with 177 features was used and the pitch factor
and reconstruction kernel were modified, it was found that
between 76 and 151 of the features were reproducible [8]. This
highlights the importance of carefully considering factors such as
the voxel size, image discretization, and the reconstruction kernel
when analyzing CT-based datasets.
PET Scans
Many studies have been conducted to assess the reproducibility of
radiomic features in PET scans, but most of them only examine
the impact of variability in scanner and imaging parameters and
do not provide specific methods for achieving reproducible fea-
tures. The full-width half maximum (FWHM) of the Gaussian
filter is the most frequently investigated reconstruction factor in
this context [1].
An important study evaluates the impact of various image
reconstruction settings in PET/CT scans using data from a phan-
tom and a patient dataset from two different scanners [9]. The
study grouped the radiomic features into intensity-based,
geometry- based, and texture-based features, and their
7 Harmonization in the Features Domain 149
MRI Sequences
The impact of test-retest, acquisition, and reconstruction settings
in MRI has been explored less extensively than for PET and CT. A
recent study investigated the robustness of radiomic features
across different MRI scanning protocols and scanners using a
phantom [11]. The results showed that the robustness of the fea-
tures varied depending on the feature, with most intensity-based
and gray-level co-occurrence matrix (GLCM) features showing
intermediate or small variation, while most neighborhood gray-
tone difference (NGTD) features showed high variation. In the
GLCM features, variance, cluster shade, cluster tendency, and
cluster prominence had poor robustness. However, these features
had high reproducibility if the scanning parameters were kept the
same, making them useful for intrascanner studies. Nevertheless,
the study had limitations, including the effect of subject move-
ment and uncertainty in lesion segmentation.
Multi-scanner Reproducibility
Multi-machine reproducibility studies involve measuring the
same image on different scanners. A recent study based on the
reproducibility of radiomic features across several MRI scanners
and acquisition protocol parameters using both phantom and
patient data with a test-retest strategy, revealed very little differ-
ences in the variability between filtering and normalizing effect
which were used for preprocessing [11]. Moreover, the intra-class
correlation coefficient (ICC) measurements showed higher repro-
ducibility for the phantom data than for the patient data, however,
the study was unable to mitigate the impact of patient’s move-
ments despite simulating movement during scanning. A similar
study also extracted stable MRI radiomic features with a mini-
mum CCC of 0.85 between data derived from 61 patients’ test and
retest apparent diffusion coefficient (ADC) maps across various
MRI systems, tissues, and vendors [14].
7.3.2 ComBat
Fig. 7.2 Application of ComBat for two radiomic features extracted from
NSCLC lesions to correct for manufacturer’s batch effect: median gray-level
intensity (top) and entropy values (bottom). On the left, original density dis-
tributions and, on the right, density distributions after harmonization with
ComBat
S. Tanadini-Lang, H. Alkadhi, y B. Baessler, “Radiomics in medical imaging—“how-to” guide and critical reflection”, Insights Imag-
ing, vol. 11, n.o 1, p. 91, dic. 2020, doi: https://doi.org/10.1186/s13244-020-00887-2. Licensed under a Creative Commons Attribution
4.0 International License https://creativecommons.org/licenses/by/4.0/
Image modality Image acquisition Reconstruction parameters Segmentation Post-processing Feature extraction
MRI • Field strength Reconstructed matrix size Manual 2D Image interpolation Mathematical formula
• Sequence design Reconstruction technique Manual 3D Intensity discretization Package
• Acquired matrix size Semi-automated Normalization
• Field of view 2D
• Slice thickness Semi-automated
• Acceleration techniques 3D
• Vendor Automated 2D
• Contrast timing Automated 3D
• Movement Size of the ROI
CT • Tube voltage Reconstruction matrix
• Milliamperage Slice thickness
• Pitch Reconstruction kernel
• Field of view/pixel spacing Reconstruction technique
• Slice thickness
• Acquisition mode
• Vendor
• Contrast timing
• Movement
PET • Field of view/pixel spacing Reconstruction matrix
• Slice thickness Slice thickness
• Injected activity Reconstruction technique
• Acquisition time Attenuation correction
• Scan timing
• Duty cycle
• Vendor
J. Lozano-Montoya and A. Jimenez-Pastor
• Movement
7 Harmonization in the Features Domain 163
References
1. van Timmeren JE, Cester D, Tanadini-Lang S, Alkadhi H, Baessler B
(2020) Radiomics in medical imaging—“how-to” guide and critical
reflection. Insights Imaging 11(1):91. https://doi.org/10.1186/s13244-
020-00887-2
2. Mühlberg A et al (2020) The technome—a predictive internal calibration
approach for quantitative imaging biomarker research. Sci Rep
10(1):1103. https://doi.org/10.1038/s41598-019-57325-7
3. van Timmeren JE et al (2016) Test–retest data for radiomics feature sta-
bility analysis: generalizable or study-specific? Tomography 2(4):361–
365. https://doi.org/10.18383/j.tom.2016.00208
4. Kessler LG et al (2015) The emerging science of quantitative imaging
biomarkers terminology and definitions for scientific studies and regula-
tory submissions. Stat Methods Med Res 24(1):9–26. https://doi.
org/10.1177/0962280214537333
5. Park JE, Park SY, Kim HJ, Kim HS (2019) Reproducibility and generaliz-
ability in radiomics modeling: possible strategies in radiologic and statis-
tical perspectives. Korean J Radiol 20(7):1124. https://doi.org/10.3348/
kjr.2018.0070
6. Midya A, Chakraborty J, Gönen M, Do RKG, Simpson AL (2018)
Influence of CT acquisition and reconstruction parameters on radiomic
feature reproducibility. J Med Imaging 5(01):1. https://doi.org/10.1117/1.
JMI.5.1.011020
164 J. Lozano-Montoya and A. Jimenez-Pastor